TALIS 2018 Technical Report
TALIS 2018 Technical Report
TALIS 2018 Technical Report
PUBE
This document, as well as any data and any map included herein, are without prejudice to the status of or
sovereignty over any territory, to the delimitation of international frontiers and boundaries and to the name
of any territory, city or area.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities.
The use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem
and Israeli settlements in the West Bank under the terms of international law.
Note by Turkey
The information in this document with reference to “Cyprus” relates to the southern part of the Island.
There is no single authority representing both Turkish and Greek Cypriot people on the Island. Turkey
recognises the Turkish Republic of Northern Cyprus (TRNC). Until a lasting and equitable solution is
found within the context of the United Nations, Turkey shall preserve its position concerning the “Cyprus
issue”.
Note by all the European Union Member States of the OECD and the European Union
The Republic of Cyprus is recognised by all members of the United Nations with the exception of Turkey.
The information in this document relates to the area under the effective control of the Government of the
Republic of Cyprus.
Photo credits:
Cover © Hill Street Studios/Gettyimages
You can copy, download or print OECD content for your own use, and you can include excerpts from
OECD publications, databases and multimedia products in your own documents, presentations, blogs,
websites and teaching materials, provided that suitable acknowledgement of OECD as source and
copyright owner is given. All requests for public or commercial use and translation rights should be
submitted to [email protected]. Requests for permission to photocopy portions of this material for public
or commercial use shall be addressed directly to the Copyright Clearance Center (CCC) at
[email protected] or the Centre français d’exploitation du droit de copie (CFC) at
[email protected].
3
Acknowledgements
The OECD Teaching and Learning International Survey (TALIS) is the outcome of
collaboration among the participating countries and economies, the OECD Secretariat, the
European Commission and an international consortium led by the International Association
for the Evaluation of Educational Achievement (IEA).
The OECD recognises the significant contributions of members of the IEA Hamburg,
Germany, IEA Amsterdam, the Netherlands, as well as Statistics Canada, who authored
this report. The principal authors of the respective chapters were as follows:
Chapter 1: Steffen Knoll and Ralph Carstens
Chapter 2: John Ainley and Ralph Carstens
Chapter 3: Ralph Carstens
Chapter 4: Michelle Djekić, David Ebbs, Sandra Dohr, Jan-Philipp Wagner, Steffen
Knoll and Viktoria Böhm
Chapter 5: Jean Dumais and Yves Morin
Chapter 6: Viktoria Böhm, Alena Becker and Christine Busch
Chapter 7: Michelle Djekić, David Ebbs, Sandra Dohr, Jan-Philipp Wagner,
Viktoria Böhm, Alena Becker and Steffen Knoll
Chapter 8: Alena Becker and Christine Busch
Chapter 9: Jean Dumais and Yves Morin
Chapter 10: Jean Dumais, Alena Becker, Jan-Philipp Wagner, David Ebbs and
Steffen Knoll
Chapter 11: Agnes Stancel-Piątak, Justin Wild, Minge Chen, Mojca Rozman,
Plamen Mirazchiyski and Hynek Cigler
Chapter 12: Ann-Kristin Koop, Falk Brese and Agnes Stancel-Piątak.
Paula Wagemaker copy-edited the report on behalf of the IEA. OECD Directorate for
Education and Skills (EDU) analysts Pablo Fraser, Gabor Fulöp and Noémie Le Donné
provided additional editorial support. Timothy L. Kennel, Bart Meuleman and
Christian Monseur provided valuable comments during the development of some chapters.
Emily Groves (EDU) provided additional proofreading and managed the layout process,
with support from Eleonore Morena, who verified the final layout, and from Henri Pearson
(EDU communications team). Emily Groves also provided administrative assistance.
We would like to gratefully acknowledge the contribution to TALIS of the late Fons van
de Vijver, who was Chair of the Technical Advisory Group and an advisor to TALIS since
the first cycle in 2008.
Table of contents
Acknowledgements ................................................................................................................................ 3
Abbreviations and acronyms .............................................................................................................. 13
Three-digit operational codes of TALIS 2018 participants ............................................................. 15
Executive summary ............................................................................................................................. 17
Governance, management and structure ............................................................................................ 17
Participants, international options and survey phases ........................................................................ 18
Framework and questionnaire development, translation ................................................................... 18
Sampling of schools and teachers ...................................................................................................... 19
Main survey administration and quality observations ....................................................................... 19
Data collection, processing, weighting and adjudication ................................................................... 20
Scaling, analysis and results table production ................................................................................... 20
Differences between TALIS 2018 and the 2008 and/or 2013 cycles................................................. 21
Chapter 1. Introduction ...................................................................................................................... 22
1.1. TALIS in brief ............................................................................................................................ 23
1.2. Participating countries and economies........................................................................................ 24
1.3. Managing the survey internationally .......................................................................................... 26
1.4. Working with national project managers .................................................................................... 27
1.5. Standardised procedures ............................................................................................................. 29
1.6. Key survey phases ...................................................................................................................... 30
References.......................................................................................................................................... 33
Notes .................................................................................................................................................. 33
Chapter 2. Development of the conceptual framework ................................................................... 34
2.1. Introduction................................................................................................................................. 35
2.2. General purpose of TALIS and policy focus for 2018................................................................ 35
2.3. Developing and refining the conceptual framework ................................................................... 37
2.4. Connecting the conceptual framework with proposed analyses ................................................. 46
References.......................................................................................................................................... 48
Notes .................................................................................................................................................. 49
Chapter 3. Development of the teacher and principal questionnaires ........................................... 50
3.1. General aims and principles ........................................................................................................ 51
3.2. Timeline ...................................................................................................................................... 54
3.3. Questionnaire Expert Group (QEG) role, membership and collaboration .................................. 56
3.4. Pilot phase ................................................................................................................................... 58
3.5. Field trial phase ........................................................................................................................... 62
3.6. Main survey phase ...................................................................................................................... 68
References.......................................................................................................................................... 78
Notes .................................................................................................................................................. 79
Chapter 4. Preparation of national survey instruments .................................................................. 81
4.1. Overview of adaptation and translation verification ................................................................... 82
4.2. Instruments requiring adaptation and translation ........................................................................ 83
Note.................................................................................................................................................. 155
Chapter 9. Estimation weights, participation rates and sampling error ...................................... 156
9.1. Overview ................................................................................................................................... 157
9.2. Estimation weights .................................................................................................................... 157
9.3. Weights for school and principal data....................................................................................... 158
9.4. Weights for teacher data ........................................................................................................... 160
9.5. Weights for the TALIS-PISA link data..................................................................................... 163
9.6. Participation rates ..................................................................................................................... 163
9.7. Sampling error with balanced repeated replication (BRR) ....................................................... 165
References........................................................................................................................................ 175
Notes ................................................................................................................................................ 175
Chapter 10. Data adjudication ......................................................................................................... 176
10.1. The general adjudication process in TALIS ........................................................................... 177
10.2. What was adjudicated in TALIS? ........................................................................................... 177
10.3. The criteria for assessment...................................................................................................... 179
10.4. Recommended usage ratings for participants ......................................................................... 179
10.5. Adjudicating the TALIS-PISA samples.................................................................................. 188
References........................................................................................................................................ 190
Notes ................................................................................................................................................ 190
Chapter 11. Validation of scales and construction of scale scores ................................................ 191
11.1. Overview ................................................................................................................................. 192
11.2. Computation of simple indices ............................................................................................... 192
11.3. Scaling procedures .................................................................................................................. 198
11.4. Results from scales evaluation and scale score creation ......................................................... 217
References........................................................................................................................................ 431
Notes ................................................................................................................................................ 433
Chapter 12. Table production and verification, analyses .............................................................. 437
12.1. Introduction ............................................................................................................................. 438
12.2. Responsibilities ....................................................................................................................... 438
12.3. Populating the table shells: General procedures ..................................................................... 438
12.4. Estimating standard errors using weights ............................................................................... 440
12.5. OECD average, TALIS average and EU total ........................................................................ 440
12.6. Estimating percentiles ............................................................................................................. 441
12.7. Use of weights in tables featuring analyses of teachers’ and principals’ data ........................ 441
12.8. Calculating parameters for the analyses of change over time ................................................. 442
12.9. Tables based on the results of regression analysis .................................................................. 443
12.10. Handling of filter-dependent questions ................................................................................. 444
12.11. Annotation rules related to data quality requirements (minimum number of cases, item
non-response) ................................................................................................................................... 444
12.12. Quality assurance and table verification ............................................................................... 445
References........................................................................................................................................ 447
Notes ................................................................................................................................................ 447
Annex A. Consortium, experts and consultants.............................................................................. 448
IEA Hamburg (Hamburg, Germany) ............................................................................................... 448
IEA Amsterdam (Amsterdam, the Netherlands) .............................................................................. 450
Tables
Table 1.1. TALIS 2018 participants in ISCED level 2 (core survey), ISCED levels 1 and 3 options
and the TALIS-PISA link .............................................................................................................. 25
Table 2.1. Country priority ratings of themes for inclusion in TALIS 2018 ISCED level 2................. 39
Table 2.2. Themes for inclusion in TALIS 2018 with frequently nominated indicators from TALIS
2013 ............................................................................................................................................... 40
Table 2.3. Map of TALIS 2018 themes to policy issues ....................................................................... 42
Table 2.4. Average rating points of themes under each policy heading, all countries (ISCED 2) ........ 42
Table 3.1. Field trial questionnaire sections .......................................................................................... 66
Table 3.2. Field trial questionnaire design ............................................................................................ 67
Table 3.3. Overview of question count across TALIS 2008, 2013 and 2018 (field trial and main
survey) ........................................................................................................................................... 76
Table 4.1. Languages used in TALIS 2018 ........................................................................................... 84
Table 4.2. Excluded TALIS 2018 principal questionnaire questions (before translation verification) . 86
Table 4.3. Excluded TALIS 2018 teacher questionnaire questions (before translation verification) ... 87
Table 4.4. List of the teacher and principal target populations in TALIS 2018 by country .................. 89
Table 5.1. TALIS 2018 target and survey populations.......................................................................... 99
Table 5.2. Establishing the sample size for TALIS 2018 .................................................................... 100
Table 5.3. Illustration of systematic random sampling with PPS ........................................................ 103
Table 5.4. Overview of the ISCED level 2 samples ............................................................................ 103
Table 5.5. Overview of the ISCED level 1 samples ............................................................................ 106
Table 5.6. Overview of the ISCED level 3 samples ............................................................................ 107
Table 5.7. Overview of the TALIS-PISA link samples ....................................................................... 107
Table 6.1. Instrument preparation time ............................................................................................... 113
Table 6.2. Originally planned and actual data collection windows for ISCED levels 1, 2 and 31 ...... 117
Table 7.1. Planned distribution of school visits across international options ...................................... 126
Table 7.2. Comparison of the planned and realised school visit design .............................................. 127
Table 7.3. IQO-reported interview responses of the SCs to questions about teacher co-operation
and comfort when participating in TALIS .................................................................................. 128
Table 7.4. IQO-reported interview responses of the SCs to questions about their experiences,
training and initial survey preparation ......................................................................................... 129
Table 7.5. IQO-reported interview responses of the SCs about explaining TALIS to participants .... 130
Table 7.6. IQO-reported interview responses of the SCs about distribution procedures .................... 130
Table 7.7. IQO-reported interview responses of the SCs about distribution time ............................... 131
Table 7.8. IQO-reported interview responses of the SCs to questions about security and anonymity 131
Table 7.9. IQO-reported interview responses of the SCs to questions about their general
impressions of the survey administration .................................................................................... 132
Table 9.1. Example of BRR-ready sample design and random assignment of pseudo PSUs ............. 166
Table 9.2. Estimated design effects for selected scales, ISCED level 2, principal data ...................... 170
Table 9.3. Estimated design effects for selected scales, ISCED level 1, principal data ...................... 172
Table 9.4. Estimated design effects for selected scales, ISCED level 3, principal data ...................... 172
Table 9.5. Estimated design effects for selected scales, ISCED level 2, teacher data......................... 172
Table 9.6. Estimated design effects for selected scales, ISCED level 1, teacher data......................... 174
Table 9.7. Estimated design effects for selected scales, ISCED level 3, teacher data......................... 174
Table 10.1. Adjudication rules for school or principal data in TALIS 2018 ....................................... 179
Table 10.2. Adjudication rules for teacher data in TALIS 2018 ......................................................... 180
Table 10.3. ISCED level 1: Principals’ participation and recommended ratings ................................ 181
Table 10.4. ISCED level 1: Teachers’ participation and recommended ratings ................................. 182
Table 10.5. ISCED level 2: Principals’ participation and recommended ratings ................................ 183
Table 10.6. ISCED level 2: Teachers’ participation and recommended ratings ................................. 185
Table 10.7. ISCED level 3: Principals’ participation and recommended ratings ................................ 187
Table 10.8. ISCED level 3: Teachers’ participation and recommended ratings ................................. 188
Table 10.9. TALIS-PISA Link: Principal's participation and recommended ratings .......................... 189
Table 10.10. TALIS-PISA Link: Teacher's participation and recommended ratings .......................... 189
Table 11.1. Measured items for school autonomy............................................................................... 194
Table 11.2. Measured items for school resources ............................................................................... 198
Table 11.3. Cut-offs for CFA model evaluation for TALIS 2018 ....................................................... 203
Table 11.4. Items average scale score equivalent table for the scale T3PERUT ................................ 212
Table 11.5. Scale counts of the invariance levels for both populations .............................................. 213
Table 11.6. Invariance level reached for each scale by ISCED level .................................................. 214
Table 11.7. Excluded populations from the estimation of the parameters of principal scales ............ 219
Table 11.8. Item wording for teacher motivation and perceptions scales ........................................... 220
Table 11.9. Reliability coefficients for teacher motivation and perceptions scales ............................ 220
Table 11.10. CFA model-data fits for scale T3PERUT ...................................................................... 223
Table 11.11. CFA model-data fit for scale T3SOCUT........................................................................ 225
Table 11.12. CFA model-data fit for scale T3VALP .......................................................................... 225
Table 11.13. Invariance tests results for scale T3PERUT ................................................................... 226
Table 11.14: Invariance tests results for scale T3SOCUT .................................................................. 227
Table 11.15. Invariance tests results for scale T3VALP ..................................................................... 227
Table 11.16. Unstandardised factor loadings for T3PERUT, T3SOCUT and T3VALP for all
countries for all populations ........................................................................................................ 227
Table 11.17. Standardised factor loadings and unstandardised intercepts for scale T3PERUT .......... 228
Table 11.18. Standardised factor loadings and unstandardised intercepts for scale T3SOCUT ......... 230
Table 11.19. Standardised factor loadings and unstandardised intercepts for scale T3VALP ............ 232
Table 11.20. Item wording for instructional practices scale items ...................................................... 234
Table 11.21. Reliability coefficients for instructional practices scales ............................................... 235
Table 11.22. CFA model-data fit for scale T3CLAIN ........................................................................ 237
Table 11.23. CFA model-data fit for scale T3COGAC ....................................................................... 239
Table 11.24. CFA model-data fit for scale T3CLASM ....................................................................... 241
Table 11.25. Invariance test results for scale T3CLAIN ..................................................................... 243
Table 11.26. Invariance test results for scale T3COGAC ................................................................... 243
Table 11.27. Invariance test results for scale T3CLASM ................................................................... 244
Table 11.28. Unstandardised factor loadings for T3CLAIN and T3COGAC for all countries for all
populations .................................................................................................................................. 244
Table 11.29. Standardised factor loadings and unstandardised intercepts for scale T3CLAIN .......... 245
Table 11.30. Standardised factor loadings and unstandardised intercepts for scale T3COGAC ........ 247
Table 11.31. Standardised factor loadings and unstandardised intercepts for scale T3CLASM ......... 249
Table 11.32. Item wording for professional practices scale ................................................................ 252
Table 11.33. Reliability coefficients for professional practices scales ................................................ 252
Table 11.34. CFA model-data fit for scale T3EXCH .......................................................................... 254
Table 11.35. CFA model-data fit for scale T3COLES ........................................................................ 256
Table 11.36. Invariance test results for scale T3EXCH ...................................................................... 258
Table 11.37. Invariance test results for scale T3COLES .................................................................... 258
Table 11.38. Unstandardised factor loadings for T3COLES for all countries for all populations ...... 259
Table 11.39. Standardised factor loadings and unstandardised intercepts for scale T3EXCH ........... 259
Table 11.40. Standardised factor loadings and unstandardised intercepts for scale T3COLES .......... 261
Table 11.41. Item wording for feedback and development scales ...................................................... 265
Table 11.42. Reliability coefficients for feedback and development scales ........................................ 265
Table 11.43. CFA model-data fit for scale T3EFFPD ......................................................................... 267
Table 11.44. CFA model-data fit for scale T3PDPED ........................................................................ 269
Table 11.45. CFA model-data fit for scale T3PDIV ........................................................................... 271
Table 11.46. CFA model-data fit for the scale T3PDBAR.................................................................. 271
Table 11.47. Invariance test results for scale T3EFFPD ..................................................................... 273
Table 11.48. Invariance test results for scale T3PDPED .................................................................... 274
Table 11.49. Invariance test results for scale T3PDIV ........................................................................ 274
Table 11.50. Invariance test results for scale T3PDBAR .................................................................... 275
Table 11.51. Unstandardised factor loadings for T3PDPED and T3PDIV for all countries for all
populations .................................................................................................................................. 275
Table 11.52. Standardised factor loadings and unstandardised intercepts for scale T3EFFPD .......... 275
Table 11.53. Standardised factor loadings and unstandardised intercepts for scale T3PDPED .......... 277
Table 11.54. Standardised factor loadings and unstandardised intercepts for scale T3PDIV ............. 279
Table 11.55. Standardised factor loadings and unstandardised intercepts for scale T3PDBAR ......... 281
Table 11.56. Item wording for teacher self-efficacy scales ................................................................. 285
Table 11.57. Reliability coefficients for teacher self-efficacy scales .................................................. 285
Table 11.58. CFA model-data fit for scale T3SECLS ......................................................................... 287
Table 11.59. CFA model-data fit for scale T3SEINS.......................................................................... 289
Table 11.60. CFA model-data fit for scale T3SEENG ........................................................................ 291
Table 11.61. Invariance test results for scale T3SECLS ..................................................................... 293
Table 11.62. Invariance test results for scale T3SEINS ...................................................................... 293
Table 11.63. Invariance test results for scale T3SEENG .................................................................... 294
Table 11.64. Unstandardised factor loadings for T3SECLS, T3SEINS and T3SEENG for all
countries for all populations ........................................................................................................ 294
Table 11.65. Standardised factor loadings and unstandardised intercepts for scale T3SECLS .......... 294
Table 11.66. Standardised factor loadings and unstandardised intercepts for scale T3SEINS ........... 296
Table 11.67. Standardised factor loadings and unstandardised intercepts for scale T3SEENG ......... 298
Table 11.68. Item wording for job satisfaction scales ......................................................................... 302
Table 11.69. Reliability coefficients for job satisfaction scales .......................................................... 302
Table 11.70. CFA model-data fit for scale T3JSENV ......................................................................... 304
Table 11.71. CFA model-data fit for scale T3JSPRO ......................................................................... 306
Table 11.72. CFA model-data fit for scale T3SATAT ........................................................................ 308
Table 11.73. Invariance test results for scale T3JSENV ..................................................................... 310
Table 11.74. Invariance test results for scale T3JSPRO...................................................................... 310
Table 11.75. Invariance test results for scale T3SATAT .................................................................... 311
Table 11.76. Unstandardised factor loadings for T3JSENV, T3JSPRO and T3SATAT for all
countries for all populations ........................................................................................................ 311
Table 11.77. Standardised factor loadings and unstandardised intercepts for scale T3JSENV .......... 311
Table 11.78. Standardised factor loadings and unstandardised intercepts for scale T3JSPRO ........... 313
Table 11.79. Standardised factor loadings and unstandardised intercepts for scale T3SATAT ......... 315
Table 11.80. Item wording for workplace well-being and stress scales .............................................. 319
Table 11.81. Omega coefficients for workplace well-being and stress scales .................................... 319
Table 11.82. CFA model-data fit for scale T3WELS .......................................................................... 321
Table 11.83. CFA model-data fit for scale T3WLOAD ...................................................................... 323
Table 11.84. Invariance test results for scale T3WELS ...................................................................... 325
Table 11.85. Invariance test results for scale T3WLOAD .................................................................. 326
Table 11.86. Invariance test results for scale T3STBEH .................................................................... 326
Table 11.87. Unstandardised factor loadings for T3WELS, T3WLOAD and T3STBEH for all
countries for all populations ........................................................................................................ 326
Table 11.88. Standardised factor loadings and unstandardised intercepts for scale T3WELS ........... 327
Table 11.89. Standardised factor loadings and unstandardised intercepts for scale T3WLOAD ....... 329
Table 11.90. Standardised factor loadings and unstandardised intercepts for scale T3STBEH .......... 331
Table 11.91. Item wording for school climate scales .......................................................................... 334
Table 11.92. Omega coefficients for populations of each participating country/economy ................. 334
Table 11.93. CFA model-data fit for scale T3DISC............................................................................ 336
Table 11.94. CFA model-data fit for scale T3STUD .......................................................................... 338
Table 11.95. CFA model-data fit for scale T3STAKE ........................................................................ 340
Table 11.96. Invariance test results for scale T3DISC ........................................................................ 342
Table 11.97. Invariance test results for scale T3STUD ....................................................................... 342
Table 11.98. Invariance test results for scale T3STAKE .................................................................... 343
Table 11.99. Unstandardised factor loadings for T3DISC, T3STUD and T3STAKE for all countries
for all populations ........................................................................................................................ 343
Table 11.100. Standardised factor loadings and unstandardised intercepts for scale T3DISC ........... 343
Table 11.101. Standardised factor loadings and unstandardised intercepts for scale T3STUD .......... 345
Table 11.102. Standardised factor loadings and unstandardised intercepts for scale T3STAKE ....... 347
Table 11.103. Item wording for the team innovativeness scale .......................................................... 350
Table 11.104. Omega coefficients for the populations in each participating country/economy ......... 350
Table 11.105. CFA model-data fit for scale T3TEAM ....................................................................... 352
Table 11.106. Invariance test results for scale T3TEAM .................................................................... 354
Table 11.107. Unstandardised factor loadings and intercepts for T3TEAM for all countries for all
populations .................................................................................................................................. 354
Table 11.108. Standardised factor loadings for scale T3TEAM ......................................................... 354
Table 11.109. Item wording for equity and diversity scales................................................................ 358
Table 11.110. Reliability coefficients for the populations of participating each country/economy .... 358
Table 11.111. CFA model-data fit for scale T3SEFE ......................................................................... 360
Table 11.112. CFA model-data fit for scale T3DIVP ......................................................................... 362
Table 11.113. Invariance test results for scale T3SEFE ...................................................................... 364
Table 11.114. Invariance test results for scale T3DIVP ...................................................................... 364
Table 11.115. Unstandardised factor loadings for T3SEFE for all countries for ISCED level 2 ........ 364
Table 11.116. Standardised factor loadings and unstandardised intercepts for T3SEFE .................... 365
Table 11.117. Standardised factor loadings and unstandardised intercepts for T3DIVP .................... 367
Table 11.118. Item wording for teaching practices ............................................................................. 369
Table 11.119. Item wording for job satisfaction scales ....................................................................... 372
Table 11.120 Reliability coefficients for job satisfaction scales ......................................................... 372
Table 11.121. CFA model-data fit for scale T3PJSENV .................................................................... 374
Table 11.122. CFA model-data fit for scale T3PJSPRO ..................................................................... 376
Table 11.123. Invariance test results for scale T3PJSENV ................................................................. 378
Table 11.124. Invariance test results for scale T3PJSPRO ................................................................. 379
Table 11.125. Invariance test results for scale T3PWLOAD .............................................................. 379
Table 11.126. Standardised factor loadings and unstandardised intercepts for scale T3PJSENV ...... 379
Table 11.127. Standardised factor loadings and unstandardised intercepts for scale T3PJSPRO ....... 381
Table 11.128. Standardised factor loadings and unstandardised intercepts for the scale
T3PWLOAD ............................................................................................................................... 383
Table 11.129. Item wording for school leadership .............................................................................. 387
Table 11.130. Omega coefficients for the school leadership scales .................................................... 387
Table 11.131. CFA model-data fit for scale T3PLEADS.................................................................... 389
Table 11.132. CFA model-data fit for scale T3PLEADP.................................................................... 389
Table 11.133. Invariance test results for scale T3PLEADS ................................................................ 391
Table 11.134. Invariance test results for scale T3PLEADP ................................................................ 392
Table 11.135. Unstandardised factor loadings for school leadership scales for all participating
countries/economies for all populations ...................................................................................... 392
Table 11.136. Standardised factor loadings and unstandardised intercepts for scale T3PLEADS ..... 392
Table 11.137. Standardised factor loadings and unstandardised intercepts for scale T3PLEADP ..... 394
Table 11.138. Item wording for school climate scale items ................................................................ 399
Table 11.139. Omega coefficients for school climate scales............................................................... 399
Table 11.140. CFA model-data fit for scale T3PACAD ..................................................................... 401
Table 11.141. CFA model-data fit for scale T3PCOM ....................................................................... 402
Table 11.142. CFA model-data fit for scale T3PLACSN ................................................................... 402
Table 11.143. CFA model-data fit for scale T3PDELI ....................................................................... 403
Table 11.144. Invariance test results for scale T3PACAD .................................................................. 405
Table 11.145. Invariance test results for scale T3PCOM .................................................................... 405
Table 11.146. Invariance test results for scale T3PLACSN ................................................................ 405
Table 11.147. Invariance test results for scale T3PDELI .................................................................... 406
Table 11.148. Unstandardised factor loadings for school climate scales for all participating
countries/economies for all populations ...................................................................................... 406
Table 11.149. Standardised factor loadings and unstandardised intercepts for scale T3PACAD ....... 406
Table 11.150. Standardised factor loadings and unstandardised intercepts for scale T3PCOM ......... 408
Table 11.151. Standardised factor loadings and unstandardised intercepts for scale T3PLACSN ..... 410
Table 11.152. Standardised factor loadings and unstandardised intercepts for scale T3PDELI ......... 412
Table 11.153. Item wording for the innovation scale .......................................................................... 415
Table 11.154. Omega coefficient for the scale on innovation ............................................................. 415
Table 11.155. CFA model-data fit for scale T3PORGIN .................................................................... 417
Table 11.156. Invariance test results for scale T3PORGIN ................................................................ 419
Table 11.157. Standardised factor loadings and unstandardised intercepts for scale T3PORGIN ..... 420
Table 11.158. Item wording for equity and diversity scale ................................................................. 423
Table 11.159. Omega coefficient for the equity and diversity scale ................................................... 423
Table 11.160. CFA model-data fit for scale T3PDIVB ....................................................................... 425
Table 11.161. Invariance test results for scale T3PDIVB ................................................................... 427
Table 11.162. Unstandardised factor loadings and intercepts for scale T3DIVB for all participating
countries/economies for all populations ...................................................................................... 427
Table 11.163. Standardised factor loadings for scale T3PDIVB......................................................... 427
Table 11.164. Item wording for distributed leadership scale .............................................................. 430
Table 11.165. Item wording for diversity practices, school scale ....................................................... 430
Table 11.166. Item wording for diversity policies, school scale ......................................................... 430
Table 11.167. Item wording for equity beliefs scale ........................................................................... 430
Table 12.1. Type of professional development undertaken by principals ........................................... 439
Table 12.2. School practices related to diversity ................................................................................. 442
Table 12.3. Principals’ gender, by school characteristics.................................................................... 445
Figures
Figure 6.1. Responsibilities of national study centres and the school co-ordinator during survey
administration .............................................................................................................................. 116
Figure 8.1. Iterative data-cleaning process .......................................................................................... 146
Figure 11.1. Illustration of the midpoint of a scale’s items ................................................................. 211
Boxes
Box 12.1. Interpretation of results derived from analysis of teachers’ and principals’ data ............... 442
Box 12.2. Annotation rules relating to data-quality requirements ...................................................... 444
1. Note by Turkey: The information in this document with reference to “Cyprus” relates to the southern part of
the Island. There is no single authority representing both Turkish and Greek Cypriot people on the Island.
Turkey recognises the Turkish Republic of Northern Cyprus (TRNC). Until a lasting and equitable solution is
found within the context of the United Nations, Turkey shall preserve its position concerning the “Cyprus
issue”.
2. Note by all the European Union Member States of the OECD and the European Union: The Republic of
Cyprus is recognised by all members of the United Nations with the exception of Turkey. The information in
this document relates to the area under the effective control of the Government of the Republic of Cyprus.
Executive summary
This technical report details the steps, procedures, methodologies, standards and rules that
the Teaching and Learning International Survey (TALIS) 2018 used to collect high-quality
data. The primary purpose of the report is to support readers of the international and
subsequent thematic reports as well as users of the public international database when
interpreting results, contextualising information and using the data. A user guide
complements this technical report and provides detailed guidance and examples for
individuals using TALIS 2018 data.
The report is structured in the form of 12 chapters followed by 9 annexes.
Chapters 1 to 12 provide an overview of the conceptual framework and questionnaire
development processes, translation and translation verification, further field operations and
the quality observations of the fieldwork in countries/economies. It describes the sampling,
weighting and adjudication procedures applied, followed by the preparation and structure
of the international databases. Finally, the scaling methodology and the scaling results
produced as well as information on the production and verification of the international
result tables are included.
Annexes A to I focus on the individuals and groups who prepared and conducted the survey,
the technical standards, forms and questionnaires used, the achieved participation rates of
the various target populations and the scaling process and its results.
The following summary highlights the key operational parameters and approaches applied
in TALIS 2018 and further detailed in this technical report. The summary then details the
key changes made during TALIS 2018 in comparison to previous cycles, especially TALIS
2013.
TALIS 2018 was initiated and managed by the Organisation for Economic Co-operation
and Development (OECD) on behalf of participating countries and economies. The TALIS
Governing Board (TGB) and its Strategic Development Group (SDG) were the primary
decision-making bodies.
The TALIS 2018 International Consortium, led by the International Association for the
Evaluation of Educational Achievement (IEA), with its premises in Amsterdam,
the Netherlands, and Hamburg, Germany, was responsible for implementing the survey at
the international level on behalf of the OECD Secretariat. The IEA’s partners were
Statistics Canada, based in Ottawa, Canada, and the Australian Council for Educational
Research (ACER) based in Melbourne, Australia.
The TALIS International Study Centre (ISC) is located at the IEA Hamburg, Germany, and
led by Steffen Knoll (IEA).
The Questionnaire Expert Group (QEG), chaired by Ralph Carstens (IEA), developed the
principal and teacher questionnaires and contributed (along with others) to the TALIS
analysis plan.
The Technical Advisory Group (TAG), chaired by Fons van de Vijver, Tilburg University,
the Netherlands, provided on-demand and further support to the consortium and the OECD
in relation to technical, methodological and analytical matters.
Forty-eight countries and economies (jointly referred to as “participants”) took part in the
“core” survey, that is, teachers and their principals in lower secondary level schools
according to the UNESCO International Standard Classification of Education (ISCED)
level 2. Within each, a national project manager (NPM) had primary responsibility for all
local work.
TALIS participants were offered one or more international survey options in addition to
the core. Fifteen participants administered TALIS 2018 in upper secondary schools (ISCED
level 3); 11 participants conducted the survey at the primary level (ISCED level 1); and 9
participants administered it in schools selected for the Programme for International Student
Assessment (PISA) 2018.
The survey was organised and conducted in three phases: (1) the pilot phase, the purpose
of which was to develop and trial the content of the survey questions with a small number
of TALIS participants; (2) the field trial phase, designed to test and evaluate the
questionnaires and item formats as well as the survey procedures and data collection
modes; and (3) the main survey phase, focused on collecting the TALIS data in the
48 participating countries and economies.
The TALIS 2018 conceptual framework guided the theory-driven development of the
survey’s school principal and teacher instruments.
Instruments reflected the participating countries/economies’ interest in monitoring changes
(trend data) over time, improving some instrument materials and introducing new materials
for topics of emerging importance.
A core and extended Questionnaire Expert Group (QEG) translated goals and priorities into
questionnaire materials and an analysis plan.
Development took place between September 2015 and August 2017 in three phases: the
pilot, which used a focus group approach; the quantitative field trial; and extensive
psychometric analyses prior to the main survey.
A series of experiments embedded at the field trial stage were analysed through use of a
system of three different, partially overlapping forms of the questionnaire for teachers.
The resulting master questionnaire for principals and teachers was used for the ISCED
level 2 core survey and, with some adaptation, for the ISCED levels 1 and 3 and the
TALIS-PISA link international options. Each questionnaire took respondents about 45 to
60 minutes to complete (English-language version).
All participating countries/economies were responsible for producing their own national
survey instruments. National study centres used the international versions (English and
French) of the survey instruments as the basis of their national questionnaires and used the
national adaptation form to document any adaptations they made to the instruments to suit
their respective national contexts.
To ensure high quality and comparability, the national instruments underwent
three verification steps: national adaptation verification, translation verification and layout
verification. Instruments were adapted, translated and verified for 48 countries/economies
for a total of 83 samples across all ISCED levels and options and 62 language versions.
Nine participants administered the survey in more than one language.
The survey activities questionnaire (SAQ) showed that most of the survey’s national project
managers (NPMs) experienced no difficulties translating the source questionnaires into
national languages or adapting them to local contexts.
TALIS 2018’s canonical sampling design was unchanged from earlier cycles, in
accordance with the OECD terms of reference (TOR) of this survey. A first-stage random
sample of 200 schools was followed by a second-stage random sample of 20 teachers from
the selected schools.
All anticipated adaptations to national conditions (e.g. number of schools or teachers in the
population of interest; excluded areas; analytical or reporting requirements) were
implemented.
The samples of schools for each participating country/economy were drawn centrally, after
which the national teams used consortium-provided software to select the samples of
teachers.
Main survey data collection took place between September and December 2017 for
Southern Hemisphere participants (with some countries extending into January 2018 as an
exception) and March to May 2018 for Northern Hemisphere participants (with some
participants starting early in February and some extending into July 2018). Extensions to
the data collection period were needed in 17 countries and economies to achieve the
required participation rates.
All but three participants administered the survey on line, the default mode.
Eleven participants administered TALIS in a mixed mode, that is, online and paper
instruments. During the main survey, 91.3% of the respondents completed the survey
on line and 8.7% completed it on paper.
Three different quality observation activities were implemented: an international quality
observation (IQO) programme overseen by the IEA Amsterdam (main survey only); a
national quality control programme run by the NPMs (field trial and main survey); and an
online survey activities questionnaire, used to gather national study centres’ feedback on
the different survey activities (field trial and main survey).
For the international quality control programme, the IEA Amsterdam recruited independent
observers who monitored the administration of the survey in a subsample of the
participating schools.
The findings from the international quality control programme generally confirmed that all
participating countries/economies implemented the procedures related to the survey
administration in accordance with technical standards.
Information from the survey activities questionnaire showed that collaboration between the
national study centres and the TALIS 2018 Consortium facilitated the successful
administration of TALIS 2018.
The IEA Hamburg applied a vast number of checks and cleaning routines to the data and
continuously sent data to the partners, the analysis teams and the OECD.
Countries/economies were provided with regular data updates for verification.
The average participation rate for ISCED level 2 principal/school data before replacement
was 85.9% (91.6% after replacement); rates ranged from 49.9% to 100.0% before
replacement. The average overall ISCED level 2 teacher participation rate was 84.3%, with
the rates ranging from 52% to 99.9%. The average overall participation rates of teachers
and principals in countries/economies that also participated in TALIS 2008 or 2013 did not
vary from the participation rates in 2018 although some countries showed improvements
while others experienced drops.
Adjudication was performed to determine the overall quality of the data, that is, whether
the data released to participants and OECD were “fit for use” as intended under the lead of
the sampling referee for TALIS 2018. The design and application of the quality assurance
processes drew on expert advice and opinion, on qualitative information and learned
judgement, and on quantitative information.
Each individual data set from all countries and economies was examined by the consortium
and arising issues were discussed and clarified with other actors as necessary. The
consortium considered the overall quality of the survey implementation and the data
yielded to be high.
The adjudication resulted in recommendations for data users regarding the quality of the
collected teacher data (samples rated as 59 good, 11 fair, 3 poor, 2 insufficient) and
principal data (samples rated as 57 good, 15 fair, 0 poor, 3 insufficient) in ISCED levels 1,
2 and 3. Recommended ratings were based on the participation rates (weighted or
unweighted) most favourable to the participants.
Two versions of the international database were created: a public-use file (PUF) available
on the OECD’s website for secondary data users and researchers; and a restricted-use file
(RUF) available only to accredited members nominated by the respective national TALIS
Governing Board members and who also accepted the respective confidentiality and
embargo rules.
Indices created in TALIS 2018 encompassed simple indices (e.g. ratios) and complex
indices (i.e. scale scores).
The scaling procedure for the complex indices was conducted by the IEA Hamburg within
the framework of multiple-group confirmatory factor analysis (MGCFA) and consisted of
two major steps – scale evaluation and scale score computation.
Differences between TALIS 2018 and the 2008 and/or 2013 cycles
The following aspects describe key technical changes applied in TALIS 2018 since 2013
especially and should be read as initial guidance to users of the public-use files until the
user guide is published later in 2019.
In keeping with the TALIS terms of reference for TALIS 2018, the sampling team
introduced a provision to control for possible shifts in coverage due to the evolution
of ISCED level definitions (ISCED-2011, previously ISCED-97). None of the
participating countries/economies reported changes to their mapping to ISCED
levels that could have adversely affected comparisons of the TALIS 2018 results
with the results of the previous two cycles.
Principal/school data were adjudicated on their own in 2018, an occurrence that
resulted in the notion of a “participating school for principal/school data” being
introduced. A school was considered “participating” if its principal returned his or
her questionnaire with at least one valid response. For the teacher data, the
minimum of 50% teacher participation remained the criterion for determining
whether a school was “participating” or not. Consequently, and in contrast to
TALIS 2008 and 2013, a school record remained on the school file if the principal
responded to the questionnaire, even if fewer than 50% of the teachers in the school
participated in the survey.
The scale score estimation implemented in TALIS 2018 was, on the one hand, more
rigorous than in TALIS 2013 and, on the other hand, more tailored to each
education system than in previous TALIS cycles. The scale scores for the 2018
cycle reflect the level of comparability across countries/economies as well as
between different ISCED levels within countries/economies. The applied
procedure enhances cross-country comparability by reducing the bias while
providing additional opportunities for comparisons between ISCED levels for
national purposes.
In this TALIS cycle, the tables for the international report were produced by the IEA and
verified by experts at the Australian Council for Educational Research (ACER) in
Melbourne, Australia.
Chapter 1. Introduction
This chapter provides an overview of the survey components and the preparation and
implementation of the Teaching and Learning International Survey (TALIS) 2018, initiated
and co-ordinated by the Organisation for Economic Co-operation and Development
(OECD). It introduces the management structure established to administer this third cycle
of the survey, the survey’s target populations and educational levels investigated, the role
of the national project managers, the key development phases of the study and the
standardised procedures implemented to allow the collection of high-quality data in 48
countries and economies.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
This OECD Teaching and Learning International Survey (TALIS) 2018 Technical Report
documents the development of the TALIS survey instruments as well as the methods used
to conduct the following: sampling; data collection, weighting, scaling and analysis; and
production of tables. It enables readers to review and understand these procedures and to
gain insight into the rigorous quality control programme that encompassed all phases of the
survey.
TALIS, the first international series of surveys to focus on the learning environment and
the working conditions of teachers in schools, offers teachers and school principals the
opportunity to provide their perspectives on school contexts. Countries can then use this
information to deepen analysis of the questions TALIS examines and to aid the
development of policy relating to these matters. TALIS data also allow countries to identify
other countries facing similar challenges and to learn from their approaches to policy
development.
The first cycle of TALIS, conducted in 2008, involved 24 countries and economies.1 The
success of this cycle, especially its valuable contribution to teacher policy development in
those countries, led to the TALIS Board of Participating Countries (BPC) and the OECD
Secretariat agreeing to conduct a second cycle, TALIS 2013.
TALIS 2013 included 34 countries and economies.2 Four additional countries and
economies3 decided to join this second TALIS cycle on a shifted schedule, which meant
they collected their main survey data in 2014, one year after the main group of participants
completed their main data collection. To ensure comparability, the four additional
participants followed the same rules, standards and principles employed for the main group.
They also used the same manuals, forms and materials.
From the time of its first cycle (2008), TALIS has required all participants to conduct its
“core” survey at the lower secondary level of education, that is, level 2 of the International
Standard Classification of Education (UNESCO-UIS, 2006[1]). Since 2013, countries have
also been able to elect to administer the survey at ISCED level 1 (primary education) and/or
ISCED level 3 (upper secondary education). A third option during TALIS 2013 invited
countries that took part in the OECD 2012 Programme for International Student
Assessment (PISA) to implement TALIS in the same schools that participated in PISA.
This option was called the “TALIS-PISA link”. Six of the 2013 countries/economies
conducted the ISCED level 1 survey, 10 the ISCED level 2 option, and 8 the TALIS-PISA
link option.
After the two successful TALIS cycles in 2008 and 2013, the TALIS BPC, the decision-
making body for the TALIS programme, now transformed into the TALIS Governing
Board (TGB), decided to conduct the third cycle of TALIS (TALIS 2018) between 2015
and 2019, with a main data collection between September and December 2017 for Southern
Hemisphere participants and between February and May 2018 for Northern Hemisphere
participants.
In 2018, 48 countries/economies participated in TALIS. As in the previous two cycles, the
“core” populations were ISCED level 2 teachers and their school leaders, that is, teachers
and leaders at the lower secondary level of school education (UNESCO-UIS, 2012[2]).4 In
addition, countries could again elect to administer the survey at ISCED level 1 and/or
ISCED level 3. Because the OECD PISA 2018 and TALIS 2018 were administered at the
same time in 2017/18, those countries/economies that took part in both programmes again
had a third option, that of implementing TALIS in the same schools that participated in
PISA. This option was again described as the TALIS-PISA link. Fifteen participants used
the ISCED level 1 option, 11 the ISCED level 3 option, and 9 the TALIS-PISA link option.
The themes and topics covered in TALIS 2018 include not only those addressed in previous
cycles but also new topics (e.g. diversity, innovation, well-being). In particular, TALIS
2018 addressed the following 11 themes and priorities related to professional
characteristics and pedagogical practices at institutional and individual levels:
teachers’ instructional practices
school leadership
teachers’ professional practices
teacher education and initial preparation
teacher feedback and development
school climate
job satisfaction (including motivation)
teacher human resource measures and stakeholder relations
teacher self-efficacy
innovation
equity and diversity.
The TALIS 2018 conceptual framework, which guided the survey’s development and also
the TALIS Consortium’s analysis plan and the OECD’s reporting plan, was released as
OECD Education Working Papers, No. 187 (Ainley and Carstens, 2018[3]). It includes
more detailed illustrations, covers priority areas and describes the mapping between these
and the initial policy aspects driving TALIS to date. It also explores the limitations evident
in such work. We strongly recommend readers of this technical report to review key aspects
of the conceptual framework as well.
Because of the positive experience in TALIS 2008 and 2013 and in recognition of the
growing number of TALIS participants interested in and capable of collecting data on line,
the OECD Secretariat decided, in 2018, to make the online mode of data collection the
default mode. Forty-five of the 48 participating countries/economies decided to collect
TALIS data predominantly on line. Only three participants opted for the paper-and-pencil-
only data collection mode. In 2013, 27 of the 34 participants collected their data on line.
The remainder of this chapter briefly describes the management of TALIS 2018 at the
international and national levels and outlines the survey’s three major phases and
milestones.
The following table (Table 1.1) lists the TALIS 2018 participants for all ISCED levels and
survey options.
Table 1.1. TALIS 2018 participants in ISCED level 2 (core survey), ISCED levels 1 and 3
options and the TALIS-PISA link
Participating country/economy ISCED level 2 (core) ISCED level 1 ISCED level 3 TALIS-PISA link
Alberta (Canada) ● ●
Australia ● ● ●
Austria ●
Belgium ●
Flemish Community (Belgium) ● ●
Brazil ● ●
Bulgaria ●
Ciudad Autónoma de Buenos ● ● ●
Aires (Argentina)
Chile ●
Colombia ● ●
Croatia ● ●
Cyprus1,2 ●
Czech Republic ● ●
Denmark ● ● ● ●
England (United Kingdom) ● ●
Estonia ●
Finland ●
France ● ●
Georgia ● ●
Hungary ●
Iceland ●
Israel ●
Italy ●
Japan ● ●
Kazakhstan ●
Korea ● ●
Latvia ●
Lithuania ●
Malta ● ●
Mexico ●
Netherlands ● ●
New Zealand ●
Norway ●
Portugal ● ●
Romania ●
Russian Federation ●
Saudi Arabia ●
Shanghai (China) ●
Singapore ●
Slovak Republic ●
Slovenia ● ●
South Africa ●
Spain ● ●
Sweden ● ● ●
Chinese Taipei ● ● ●
Participating country/economy ISCED level 2 (core) ISCED level 1 ISCED level 3 TALIS-PISA link
Turkey ● ● ● ●
United Arab Emirates ● ● ●
United States ●
Viet Nam ● ● ● ●
1. Note by Turkey: The information in this document with reference to “Cyprus” relates to the southern part of the Island.
There is no single authority representing both Turkish and Greek Cypriot people on the Island. Turkey recognises the
Turkish Republic of Northern Cyprus (TRNC). Until a lasting and equitable solution is found within the context of the
United Nations, Turkey shall preserve its position concerning the “Cyprus issue”.
2. Note by all the European Union Member States of the OECD and the European Union: The Republic of Cyprus is
recognised by all members of the United Nations with the exception of Turkey. The information in this document relates
to the area under the effective control of the Government of the Republic of Cyprus.
In January 2016, the OECD entered a partnership with the International Association for the
Evaluation of Educational Achievement (IEA) and its consortium partners, Statistics
Canada (Ottawa, Canada) and the Australian Council for Educational Research (ACER,
Melbourne, Australia). Under this partnership, the OECD commissioned the IEA Hamburg
office as the international study centre (ISC) to conduct TALIS 2018. The consortium
included staff from both IEA offices in Amsterdam and Hamburg and from Statistics
Canada and ACER.
The team at the ISC was led by Steffen Knoll (co-director for operations) and
Ralph Carstens (co-director for content) and included Friederike Westphal (study
co-ordinator until May 2016), Viktoria Böhm (study co-ordinator since May 2016),
Malgorzata Petersen (study co-ordinator assistant since October 2017), Juliane Kobelt
(study co-ordinator since November 2018), Alena Becker (international data manager), and
Christine Busch (deputy international data manager). Ralph Carstens (IEA Hamburg)
chaired the Questionnaire Expert Group (QEG). Its role was to co-ordinate the team of
internationally selected researchers responsible for developing the TALIS 2018
questionnaires. John Ainley (ACER), in collaboration with the QEG members, prepared
the conceptual framework. The analysis team at the IEA Hamburg, led by Agnes Stancel-
Piatak, was responsible for data scaling and table production. The team developed and
implemented the validation and scaling procedures, produced the scale scores, prepared the
datasets for analysis and produced the tables according to pre-agreed analysis and reporting
plans.
At the IEA Amsterdam, the IEA financial director Roel Burger managed the financial and
contractual affairs in co-operation with Christian Groth, head of accounting and controlling
at the IEA Hamburg. David Ebbs co-ordinated the translation verification of the national
survey questionnaires in 62 language versions. cApStAn Linguistic Quality Control,
specialising in semantic quality control of translations in a range of international surveys
and based in Brussels, Belgium, was contracted to support this work. Michelle Djekic was
responsible for managing the international quality observation programme in all
participating countries/economies by contracting, training and supervising independent
quality observers. She was supported by Sandra Dohr, who took over this responsibility in
August 2018.
The study’s sampling referee, Jean Dumais, and the sampling team manager, Yves Morin,
both of Statistics Canada, were responsible for the survey’s international sample design
and its national sampling plans, implementation, weighting and adjudication.
The OECD Secretariat’s Directorate for Education and Skills in Paris, France, responsible
for the overall supervision of the project across the participating countries/economies and
TALIS governing bodies, was led by Karine Tremblay. She was supported by Pablo Fraser
and Noémie Le Donné.
The OECD Secretariat appointed Fons van de Vijver (Tilburg University) as chair of the
TALIS 2018 Technical Advisory Group (TAG). Its task was to provide the OECD and the
TALIS Consortium and QEG with expert advice and guidance on methods, processes and
analyses.
In November 2015, the BPC initiated the process of establishing a national centre in each
participating country/economy. Each centre was to be led by an experienced national
project manager (NPM) who would be asked to adhere to consortium guidance on NPM
roles and responsibilities and would have primary responsibility for preparing and
co-ordinating the survey at the national level. The experience and expertise of the NPM
and the staff in his or her centre (e.g. involvement in other international large-scale
assessments) strongly influenced how the various centre tasks were apportioned and
managed.
The consortium also recommended that each centre appoint not only a national data
manager (NDM) to oversee and implement technical and data-related work but also a
national sampling manager (NSM) to support the work of the NPM in situations involving
complex sample designs and possible national additions and extensions to the survey. The
number of staff members in the centres varied considerably from one country/economy to
the next depending on its size and how it chose to organise the national data collection
work. Some NPMs tendered for and contracted external survey organisations to help them
conduct the study centre’s scientific and/or operational work.
The tasks required of the NPMs, data managers and/or sampling managers included the
following:
establishing an overall preparation and administration schedule in co-operation
with the ISC
attending NPM meetings in order to become familiar with all instruments, materials
and survey procedures
providing Statistics Canada with an up-to-date national sampling frame of ISCED
level 2 schools and, where applicable, ISCED levels 1, 3 and PISA schools
discussing national design options, such as oversampling, with Statistics Canada’s
sampling experts
performing within-school listing, sampling and tracking
carrying out quality control and plausibility checks on teacher lists and samples to
identify, for example, teacher lists that had been abbreviated to include only those
teachers who had agreed to participate or that were otherwise incomplete/inaccurate
liaising, with respect to the TALIS-PISA link option, with the PISA 2018 NPM to
obtain the school frames, files and other data items required to successfully
implement the link
exchange between all involved parties. All systems were operated by the ISC and, in the
interests of ensuring confidentiality and security, only authorised personnel could access
information.
During the four-year survey cycle, the consortium held four meetings for all NPMs and
NDMs, during which survey progress was reported and discussions on materials,
procedures, standards and results were held. NPMs also had bilateral and plenary
opportunities to exchange experiences and learn about approaches to, for example, coping
with survey fatigue, ensuring confidentiality and simultaneously managing the
international options and other surveys.
The TALIS 2018 technical standards (see Annex B of this current report), prepared by the
consortium and approved by the TGB, provided participants with a high level of guidance
during all stages of preparation, administration and data work. The standards encompassed
the generally agreed-upon best practices in survey research to adhere to when conducting
a project (see, for example, Biemer and Lyberg (2003[4])]; Martin, Rust, and Adams
(1999[5])]; OECD (2013[6]); and Statistics Canada (2009[7])). According to the total survey
error framework (Biemer and Lyberg, 2003[4]), adherence to these standards is key to
ensuring the validity, reliability and comparability of questionnaires and data. The
consortium also developed an extensive set of operational manuals and guidelines
describing the steps that all participants needed to take to ensure successful implementation
of the survey.
TALIS used two questionnaires to collect data: a principal questionnaire, completed by
school leaders, and a teacher questionnaire, completed by the sampled teachers.
Respondents could choose to fill in the questionnaires on line or with paper and pencil. The
ISC provided the source versions of the questionnaires in English and in French. National
study centre personnel then adapted the instruments to suit local contexts, applying
standardised adaptation rules as they did so. After the consortium had verified all
adaptations, the national centres translated the English or French source versions of the
questionnaires into the local language(s).
For the field trial stage of TALIS 2018, the national study centres used the IEA
eAssessment System to translate the source versions of the survey instruments into the local
language(s). For the main survey, participants used the IEA Online Survey System (OSS)
for this purpose (see Chapters 6 and 8). Independent translation verifiers were employed to
conduct international translation verification of all translated survey instruments. The
TALIS consortium at the IEA Amsterdam was responsible for centrally co-ordinating the
work of the translation verifiers.
The Statistics Canada team performed all school sampling and weighting procedures in line
with established standards and guidelines (more details on the sampling procedures and on
the sampling weights and participation rates appear in Chapters 5 and 9 respectively). The
TALIS NPMs used the IEA Within-School Sampling Software (WinW3S) to draw the
teacher sample for each nationally sampled school and to estimate participation rates.
The IEA Amsterdam co-ordinated the quality observation of the data collection at the
international level, while the NPMs took on this responsibility at the national level.
International quality observers (IQOs) received an intensive two-day period of training (see
Chapter 7).
Once the TALIS Consortium was confident all data quality standards had been fully met,
they recommended the data for inclusion in the TALIS 2018 database and subsequent
reports of findings. If, however, the consortium considered standards had not been fully
met, it implemented an adjudication process in consultation with the OECD Secretariat and
TAG and used it to ascertain the extent to which the data quality had been compromised.
The results determined whether data could (or could not) be recommended for
unconditional inclusion in the datasets and consequent reporting.
The TALIS design included three main phases: a qualitative pilot study conducted with a
reduced number of participants, a field trial and the main survey. Each phase included all
TALIS participants.
Based on the positive experiences in previous TALIS cycles, the consortium asked
participants to establish focus groups that included teachers and principals. The purpose of
these groups was to discuss the proposed field trial survey items and to provide feedback
on their functioning, cultural applicability and other aspects (see also Chapter 3).
The field trial and main survey followed a quantitative approach. The consortium required
all participants to run the trial according to the standardised procedures outlined in the
TALIS 2018 technical standards (see Annex B). Countries that had opted to participate in
one or more of the international options had to trial them as well.
The consortium held the first meeting of NPMs at the end of April 2016 in Lübeck,
Germany, prior to administration of the pilot study. The purpose of the meeting was to
present and discuss the survey’s draft conceptual framework, the sampling procedures, the
roadmap for instruments and materials, the planned survey operations procedures and the
overall project governance and responsibilities. NPMs were also introduced to the overall
schedule for TALIS 2018, to communication procedures and to best-practice survey
development, administration and implementation procedures collated during TALIS 2008
and 2013.
ease with which they could be fine-tuned for the field trial (with ease based on
analysis of the pilot outcomes).
In March 2016, the ISC conducted a webinar to initiate preparation of the pilot study. The
centre also released pilot guidelines to national centres.
The pilot study, conducted in May 2016, involved 11 countries from different locations and
cultural and language backgrounds. Five countries contributed to the piloting with respect
to ISCED level 1 and four with respect to ISCED level 3. NPMs established focus groups
of seven to nine members to discuss the pilot instruments. Although translating the
instruments into national languages was not required if all focus group members were
fluent in English, some countries did complete translation work.
The teacher questionnaire included a larger number of items than were needed for testing
and trialling. This number meant that a rotated design could be used later during placement
of items in the field trial instruments. Because of the large number of items during the pilot
study, each group of teachers focused on only half (A or B) of them to keep their
questionnaire completion time within acceptable limits. However, the ISC advised that
each country/economy needed to have only one principal focus group because the number
of items used in the principal questionnaire for the pilot was similar to the number in the
2013 main survey principal questionnaire.
checks. Extensive analysis of the field trial data by the IEA’s analysis team and QEG (the
core group as well as an extended set of experts), consultations with the TAG, the OECD
Secretariat and the TGB led to key improvements and changes in the main survey
instruments, released for translation in August 2017 (refer to Chapter 3 for more details).
The IEA analysis team referred to the field trial data when conducting the scaling of simple
and complex scales and preparing table production syntax for the main survey. They also
carried out detailed analyses designed to evaluate the scales across TALIS cycles and
across ISCED levels.
References
Ainley, J. and R. Carstens (2018), “Teaching and Learning International Survey (TALIS) 2018 [3]
Conceptual Framework”, OECD Working Papers, No. 187, OECD Publishing, Paris,
http://dx.doi.org/10.1787/799337c2-en.
Biemer, P. and L. Lyberg (2003), Introduction to Survey Quality, John Wiley & Sons, Hoboken, [4]
New Jersey.
Martin, M., K. Rust and R. Adams (1999), Technical Standards for IEA Studies, International [5]
Association for the Evaluation of Educational Achievement, Amsterdam.
OECD (2013), TALIS 2013 Technical Standards, OECD Publishing, Paris. [6]
UNESCO-UIS (2006), International Standard Classification of Education (ISCED) 1997, re- [1]
edition, UNESCO Institute for Statistics, Montreal,
http://uis.unesco.org/sites/default/files/documents/international-standard-classification-of-
education-1997-en_0.pdf.
Notes
1
TALIS 2008 participants: Australia, Austria, Belgium (Flemish Community), Brazil, Bulgaria,
Denmark, Estonia, Hungary, Iceland, Ireland, Italy, Korea, Lithuania, Malaysia, Malta, Mexico,
Norway, Poland, Portugal, Slovenia, the Slovak Republic, Spain, Turkey.
2
TALIS 2013 participants: Abu Dhabi (United Arab Emirates), Alberta (Canada), Australia, Brazil,
Bulgaria, Chile, Croatia, Cyprus,* the Czech Republic, Denmark, England (United Kingdom),
Estonia, Finland, Flanders (Belgium), France, Iceland, Israel, Italy, Japan, Korea, Latvia, Malaysia,
Mexico, the Netherlands, Norway, Poland, Portugal, Romania, Serbia, Singapore, the Slovak
Republic, Spain, Sweden, the United States (*See footnotes 1 and 2 in Table 1.1).
3
TALIS 2013 additional participants: Georgia, New Zealand, Russia, Shanghai (China).
4
In TALIS 2018, the updated International Standard Classification of Education (ISCED), as
published by UNESCO-UIS in 2012, was used to identify ISCED levels in schools.
This chapter describes the development of the conceptual framework that guided TALIS
2018. It describes the theoretical and policy underpinnings of the survey, articulates its
research emphases and links to existing knowledge, and sets out the indicators included in
the TALIS 2018 instruments. TALIS 2018 not only retains a focus on enduring issues
previously surveyed in TALIS 2013 and TALIS 2008 but also introduces some new and
changing aspects of teaching and learning.
2.1. Introduction
This chapter outlines the process involved in developing the conceptual framework for the
Teaching and Learning International Survey (TALIS) 2018, which was published as OECD
Education Working Papers, No. 187 (Ainley and Carstens, 2018[1]). This chapter has three
sections: TALIS’s general purpose and policy focus; knowledge relevant to the survey’s
themes and main indicators; and the design of TALIS 2018.
TALIS is relevant to this endeavour because Target 4.c of Goal 4 specifically addresses the
role of teachers in ensuring quality education: “By 2030, substantially increase the supply
of qualified teachers, including through international co-operation for teacher training in
developing countries, especially least developed countries and small island developing
States” (United Nations, 2015, p. 17[9]).
Target 4.c consists of one global indicator and six thematic indicators. The global indicator
is “Proportion of teachers in: (a) pre-primary; (b) primary; (c) lower secondary; and (d)
upper secondary education who have received at least the minimum organised teacher
training (e.g. pedagogical training) pre-service or in-service required for teaching at the
relevant level in a given country” (UNESCO, 2016, p. 81[8]).
The six thematic indicators are the pupil to trained teacher ratio by education level; the
proportion of teachers qualified according to national standards, by education level and
type of institution; the pupil to qualified teacher ratio, by education level; the average
teacher salary relative to other professions requiring a comparable level of qualification;
the teacher attrition rate, by education level; and the percentage of teachers who received
in-service training in the last 12 months, by type of training. TALIS provides data on
teacher certification and the highest educational level attained as a proxy for qualified
teachers and thereby addresses the extent to which countries have achieved Goal 4. In
addition, the TALIS indicator of professional development provides information on the
percentages of teachers who received in-service training in the last 12 months.
Participating countries regarded some themes as very high priorities (e.g. school leadership
and teachers’ instructional practices) and other themes as less important (e.g. teacher
attrition and turnover rates and the sociological composition of teachers). Substantial
cross-country variation was evident in these rankings. In general, the highest-rated themes
were those that most closely matched the countries’/economies’ priorities. For example,
one-third of countries/economies gave a relatively low rating to the theme of teachers’
professional practices.
During the second step, countries were asked to consider the 20 themes to which they had
assigned points and to state which of the 94 indicators should be assigned to each of these
themes. The third step asked countries to indicate which of the indicators used in TALIS
2013 they thought should be maintained in TALIS 2018 to permit analysis of change
between these two cycles.
Table 2.1. Country priority ratings of themes for inclusion in TALIS 2018 ISCED level 2
Source: OECD (2015[7]), “Guiding the Policy and Content Focus of TALIS 2018”, EDU/INES/TALIS(2015)
3 (internal document), Directorate for Education and Skills, OECD, Paris, pp. 14-15.
Because of the intention to lower the number of themes in TALIS 2018, the TGB decided
to proceed with no more than ten themes that, in combination, would inform all
five identified policy issues, namely, school policies supporting effectiveness, developing
teachers within the profession, effective teachers and teaching, attracting teachers to the
profession, and retaining teachers in the profession. Another decision was to place a slight
emphasis on those themes among the 20 that addressed policies related to school and
teacher effectiveness. This decision reflected the fact that themes attracting the highest
ratings were those concerned with “school policies supporting effectiveness”.
In addition to the policy rating exercise, ministers of education in the participating
countries/economies were invited to highlight the themes they considered to be key areas
of interest. During this part of the exercise, some of the initially proposed themes were
combined with others, which resulted in variations to the initial list. For example, Theme 5
incorporated the two elements of teacher feedback and teacher development from the
original list. In the end, nine themes were agreed to, even though the relative importance
accorded to each of them varied across the participating countries/economies. Table 2.2
sets out those nine themes, together with the most frequently nominated indicators from
TALIS 2013.
Innovation was initially seen as a cross-cutting issue closely related to teachers’
instructional practices and school climate. However, it emerged as an explicit theme
(Theme 10) during discussions by the TGB and QEG and was assigned these indicators:
teachers’ openness to adopting innovative practices
types of innovation in the school in the past year
types of innovation in the target classroom in the current or past school year
perceptions regarding the barriers to and incentives for the adoption of innovation
evaluation and dissemination of innovative practices in the school.
The TGB and QEG originally considered equity and diversity to be encapsulated in the
substance of each of the nine themes, but the TALIS participants and policy stakeholders
mutually decided to include this theme (Theme 11) as a theme of high contemporary
importance.
One further decision resulting from the priority exercise established that although there
might be a need to adapt specific questionnaire items to suit respondents at different levels
of education, the questionnaires for ISCED levels 1, 2 and 3 should all address the same
themes.
Table 2.2. Themes for inclusion in TALIS 2018 with frequently nominated indicators from
TALIS 2013
Theme Frequently nominated indicators
1. Teachers’ instructional a. beliefs about teaching
practices b. classroom climate in target class
c. pedagogical practices in target class
d. classroom management in target class
e. individualised/differentiated teaching (including gifted students) in target
class
f. teachers’ views regarding barriers to implementing a variety of practices
g. classroom composition and class size in target class
h. lesson time distribution in target class
Table 2.3 also shows the main connections between themes and policy areas. The
connections arise either because the theme is, by definition, part of the policy area or
because the theme encapsulates factors that could have potentially strong influences on the
policy area. The TGB and QEG in mutual consultation agreed that the themes collectively
and reasonably represent all five policy areas. Also, because there are more themes than
policy areas, more than one theme necessarily addresses the policy areas. Four of the
TALIS 2018 themes inform the policy area of effective teachers and teaching. Similarly,
some themes inform more than one policy area. For example, the theme “teacher human
resource measures and stakeholder relations” connects to attracting teachers, retaining
teachers and school effectiveness.
Source: Based on information from OECD (2015[7]), “Guiding the Policy and Content Focus of TALIS 2018”,
EDU/INES/TALIS(2015)3 (internal document), Directorate for Education and Skills, OECD, Paris.
Table 2.4 presents the average rating given to the themes within each of the five high-level
policy headings. The top-rated policy issues were school policies supporting effectiveness
and developing teachers within the profession. The very small differences in average scores
across the five policy issues indicate support for balanced coverage of all five policy issues.
Table 2.4. Average rating points of themes under each policy heading, all countries
(ISCED 2)
Source: OECD (2015[7]), “Guiding the Policy and Content Focus of TALIS 2018”, EDU/INES/TALIS(2015)
3 (internal document), Directorate for Education and Skills, OECD, Paris, p. 16.
background data were also intended to reveal basic characteristics likely to be of interest in
terms of their relationship to other indicators, as descriptive information about schools and
systems, and in providing an understanding of the contexts in which data about TALIS
themes and indicators are interpreted.
School and classroom context can be conceptualised either as the social composition of the
school and classroom or as the neighbourhood in which the school is located. Debate
continues on the extent to which the overall characteristics of the student population have
an effect on student learning outcomes after statistically allowing for the effects for
individual students (Borman and Dowling, 2010[11]). However, analyses of PISA results
suggest that, in most of the countries participating in that study, students, regardless of their
own socio-economic background, are advantaged scholastically if they attend “a school
whose students are, on average, from more advantaged socio-economic backgrounds”
(OECD, 2004, p. 189[12]).
Of concern to TALIS was the degree to which the effects of school composition on student
achievement are influenced by differences in the characteristics of teachers and differences
in approaches to teaching that are associated with differences in the composition of the
school population. Students with migrant or refugee backgrounds and their education are a
priority for many countries (OECD, 2015[7]). The TGB, therefore, suggested that TALIS
2018 should examine teaching and school practices in schools with varying percentages of
students with an immigrant background. The TGB also expressed interest in the extent to
which school structural characteristics and geographic location affect student achievement
and other outcomes, with that influence mediated by the impact these characteristics and
location have on how teaching takes place.
The TGB also indicated that TALIS 2018 should continue to include consideration of the
influence of the percentage of teachers employed on a part-time basis. A number of
countries appear to have experienced an increase in the percentages of teachers who work
part-time. Williamson, Cooper, and Baird (2015[13]) documented variations in the incidence
of part-time work across countries, with Ireland, the Netherlands, Switzerland and
the United Kingdom having relatively high rates of part-time work. The TGB considered
that TALIS 2018 provided an opportunity to investigate variations within and across
countries in the percentages of teachers employed on a part-time basis and the extent to
which these variations are associated with variations in other aspects of schooling.
PISA 2018
The fact that TALIS 2018 and PISA 2018 were implemented in the same year made it
possible to harmonise the two surveys, as did the fact that PISA includes questionnaires for
teachers and principals. Consideration was given to the possible synergies between TALIS
and PISA presented in a joint conceptual framework (OECD, 2015[14]). The TALIS and
PISA teams discussed these matters not only at the OECD Secretariat but also during
exchanges with the PISA 2018 contractors, especially the Educational Testing Service
(ETS) and the German Institute for International Educational Research (DIPF).
The alignment sought was in general terms for the TALIS-PISA link international option.
The teams agreed that although some themes (TALIS) and modules (PISA) had similarities,
they were not closely aligned. The development work for PISA 2018 had commenced in
2014, with frameworks and field trial materials finalised towards the end of 2015.
Therefore, each survey considered pragmatic arrangements for 2018 in the absence of a
fully developed link between the surveys. The areas in which materials aligned with PISA
2018 were those concerned with job satisfaction, self-efficacy and school climate. Both
surveys also included specific questions on initial teacher education and teaching in
culturally diverse settings (equity and diversity).
programmes. TALIS also collected data on the support available for new teachers, given
these teachers tend to be the ones at the greatest risk of teacher attrition (OECD, 2005[2]).
instructional practices) and cross-theme analyses (e.g. analyses of the associations between
teachers’ instructional practices and teachers’ professional practices). Some of these
analyses can be conducted at a teacher or school level while others can be conducted only
at a school or country/economy (system) level. Time-series (cross-cycle) analyses enable
monitoring of changes over time (assuming measurement equivalence over time has been
established).
The reporting plan for TALIS 2018 outlines the content and structure of the initial report
of the survey’s findings (OECD, 2018[16]). The report is planned as a policy-oriented
document designed to stimulate reflection on practice and to relate to broader research on
the TALIS 2018 themes. The initial report will be released in two volumes dealing with
two main dimensions of teachers’ and school leaders’ work. The first volume will focus on
the knowledge and skills dimension of professionalism and will examine contemporary
teaching and learning practices, as well as mechanisms available to support teachers’ and
school leaders’ learning throughout their career pathways in order to deliver quality
schooling for all students. The second volume will then focus on teachers and school
leaders’ other professional practices and examine the mechanisms available to support and
strengthen their professionalism.
Across these topics, there will be an emphasis on:
reporting of results about both teachers and school leaders
commenting on meaningful international comparisons
discussing results in context
reporting relations between themes (cross-theme analyses)
describing trends (cross-cycle analyses).
Additional thematic reports or policy briefs will cover the following options and themes:
primary education teachers and principals (ISCED level 1 option) and upper
secondary education teachers and principals (ISCED level 3 option)
schools performing against the odds (TALIS-PISA link option)
equity issues across schools, teachers and students
supporting teachers’ well-being and retention.
References
Ainley, J. and R. Carstens (2018), Teaching and Learning International Survey (TALIS) 2018 [1]
Conceptual Framework, OECD Publishing, Paris, http://dx.doi.org/10.1787/799337c2-en.
Borman, G. and M. Dowling (2010), “Schools and inequality: A multilevel analysis of [11]
Coleman’s equality of educational opportunity data”, Teachers College Record, Vol. 112/5,
pp. 1201-1246, http://www.tcrecord.org/library/abstract.asp?contentid=15664.
Kyriakides, L., C. Christoforou and C. Charalambous (2013), “What matters for student learning [10]
outcomes: A meta-analysis of studies exploring factors of effective teaching”, Teacher and
Teacher Education, Vol. 36, pp. 143-152, http://dx.doi.org/10.1016/j.tate.2013.07.010.
OECD (2018), “Proposal for the TALIS 2018 Reporting Plan”, EDU/TALIS/GB(2017)2/REV4 [16]
(internal document), Directorate for Education and Skills, OECD, Paris.
OECD (2015), “Guiding the Policy and Content Focus of TALIS 2018”, [7]
EDU/INES/TALIS(2015)3 (internal document), Directorate for Education and Skills, OECD,
Paris.
OECD (2015), “Joint Conceptual Framework for TALIS and PISA Synergies”, [14]
EDU/INES/TALIS(2015)6 (internal document), Directorate for Education and Skills, OECD,
Paris.
OECD (2014), TALIS 2013 Results: An International Perspective on Teaching and Learning, [5]
TALIS, OECD Publishing, Paris, http://dx.doi.org/10.1787/9789264196261-en.
OECD (2010), TALIS 2008 Technical Report, TALIS, OECD Publishing, Paris, [3]
https://dx.doi.org/10.1787/9789264079861-en.
OECD (2009), Creating Effective Teaching and Learning Environments: First Results from [4]
TALIS, TALIS, OECD Publishing, Paris, https://dx.doi.org/10.1787/9789264068780-en.
OECD (2005), Teachers Matter: Attracting, Developing and Retaining Effective Teachers, [2]
Education and Training Policy, OECD Publishing, Paris,
https://dx.doi.org/10.1787/9789264018044-en.
OECD (2004), Learning for Tomorrow’s World: First Results from PISA 2003, PISA, OECD [12]
Publishing, Paris, http://dx.doi.org/10.1787/9789264006416-en.
UNESCO (2016), Education 2030: Incheon Declaration and Framework for Action for the [8]
Implementation of Sustainable Development Goal 4, UNESCO, Paris,
http://uis.unesco.org/sites/default/files/documents/education-2030-incheon-framework-for-
action-implementation-of-sdg4-2016-en_2.pdf.
United Nations (2015), Transforming our World: The 2030 Agenda for Sustainable [9]
Development, United Nations, New York,
http://www.un.org/ga/search/view_doc.asp?symbol=A/RES/70/1&Lang=E.
Williamson, S., R. Cooper and M. Baird (2015), “Job-sharing among teachers: Positive, negative [13]
(and unintended) consequences”, The Economic and Labour Relations Review, Vol. 26/3,
pp. 448-464, http://dx.doi.org/10.1177/1035304615595740.
Notes
1
The QEG members working on the themes (and the themes for which they were responsible) were
S. Blömeke, R. Scherer and T. Nilsen (teachers’ instructional practices, teacher education and initial
preparation, teacher self-efficacy, innovation); D. Muijs (school leadership, teacher human resource issues
and stakeholder relations); H. Hollingsworth (teachers’ professional practices, teacher feedback and
continuing development); H. Price (school climate, job satisfaction); and F. van de Vijver and J. Ainley
(equity and diversity). D. Kaplan contributed to the QEG’s consideration of research methods across all
themes.
2
The extended QEG reviewers were E. Aller, S. Howie, M. Mok, S. Seeber and S. Taut.
The development of the survey instruments for TALIS 2018 described in this chapter was
strictly guided by the conceptual framework (see Chapter 2). A Questionnaire Expert
Group (QEG) was established under TALIS 2018 International Consortium management
to translate the identified goals and priorities into survey questionnaires, support materials
and an analysis plan. At the operational level, the development and validation of
instruments were implemented in several phases, with the observations and outcomes of
the previous step influencing revisions and plans for the subsequent phase. The key
challenges related to the significant increase in the number of participating
countries/economies, the interest in monitoring changes over time while keeping a
forward-looking approach, and the intention to embed a series of experiments at the field
trial stage.
Throughout its existence, TALIS’s main goal has been to generate valid, internationally
comparable information that is relevant to teachers and teaching and maintains an emphasis
on aspects at the system, school and teacher levels that are known or can be expected to
affect student learning. TALIS develops instruments for comparing these contexts cross-
culturally. Large, carefully selected representative samples of survey respondents and
modern methods of data collection and of quantitative analysis enable formulation of broad
inferences about the surveyed populations. Together, these approaches also allow the
development of conclusions about important relationships between and among factors of
interest within and across countries and economies.
Accordingly, TALIS 2018 strived to apply the following set of key principles to its
development process to the maximum extent possible:
a theory-driven development guided by a conceptual framework (see Chapter 2)
maximum country/economy input (at the policy as well as the operational level)
into the development and selection of constructs, questions and items
extensive consultations with experts in the survey’s corresponding thematic areas
as well as experts in questionnaire and sample design
retention of selected constructs, variables and measures from TALIS 2013 to allow
analyses of changes across TALIS cycles
improvements to questions and items as deemed appropriate and validated through
comparative experiments
descriptions of the cross-cultural validity of measures.
Use of the same data collection instruments across countries/economies, along with some
minor contextual alterations for the ISCED levels 1 and 3 options, allows TALIS to validly
document the variation in teacher practice and development that exists not only across
countries/economies and within each country/economy but also, where applicable, across
levels of education. TALIS also generates, for each country/economy and across
countries/economies with available data, reliable information about changes over time in
key aspects between indicators relating to teachers and the contexts of their work.
Based on the initial priorities voiced by the TALIS Board of Participating Countries (BPC)
in 2015 (2015[1]), the TALIS 2018 International Consortium, in close collaboration with
the Questionnaire Expert Group (QEG) members (see next section), drafted a
comprehensive development plan (OECD, 2015[2]) in late 2015 designed to capture general
considerations and proposals that would inform the start of the development work and also
establish initial proposals for each theme/topic that the BPC could consider and comment
on during its nineteenth meeting. This plan for the conceptual framework articulated the
survey’s initial research focus and direction. It also articulated the survey’s theoretical
underpinnings, existing knowledge and evidence, and the methods that would be used to
guide the development of the TALIS 2018 instruments, indicators and operations. The plan
described an iterative process in which academics and scholars formulated concepts,
discussed them with relevant stakeholders and then revised and reformulated the concepts
as necessary.
The QEG members developed for each theme (and the cross-cutting issue of innovation)
initial free-format concept notes and discussed these at their first meeting in September
2015. Members standardised the notes around four headings to the extent possible:
country priorities and other inputs
o interpreting themes within the OECD “policy and content” focus document
o interpreting aspects of themes as described in other OECD documents,
especially PISA
theoretical background and justification
o a general introduction to each issue, with particular consideration given to the
established TALIS 2013 legacy
o the general direction to be taken during the development of each theme
(e.g. new influences, paradigm shifts, incomplete coverage in 2008 and 2013)
o a brief review of relevant research literature, including studies that had
produced some evidence of possible causal relationships
o a focus on the nature of TALIS, on teaching and learning from teachers’
perspectives, and on the working conditions of teachers
o a statement on why TALIS should investigate each issue
key development directions and most important changes
o major areas of new development as well as areas needing re-working and/or a
shift in focus
o high-level descriptions of materials (questions, items, scales) to be retained,
reworked, introduced or dropped (“chopping board”)
o triangulation and possible harbouring of crosscutting issues
analytical potential and indicators
o outlook on the type of research questions for which data are being sought
o relationships to other themes and systems, schools, teacher characteristics
(e.g. the link to and interaction between school climate and aspects of
leadership)
o policy relevance and use (general, current, emerging)
o a brief sketch of potential indicators.
The initial drafts of the concept notes outlined a wide range of possible directions for
development and it soon became clear that full implementation of all changes would not be
compatible with the aim of retaining a sufficiently high proportion of (key) materials that
would enable links back to 2013.
Ensuing discussion of the development plan at the next BPC meeting, therefore, focused
on:
the need, because of the requirement to maintain an average response time of
45 minutes for the English version of the principal questionnaire (PQ) and for the
teacher questionnaire (TQ), to achieve balance across each questionnaire with
respect to the following: maintaining existing questions from TALIS 2013, revising
As indicators drill deeper, the number of questionnaire items required increases. Some
indicators collect information about whether particular features exist, such as a mentoring
system, for example. Some indicators collect slightly richer data on the scale or frequency
of certain features. Examples include the typology of mentoring offered or the frequency
with which some activities occur. Other indicators aim to collect enough information to
determine how something works, such as identifying the features of a professional
development activity that has had a perceived positive impact. Finally, some indicators
collect data on what schools, principals and teachers do, and what degree of impact can be
attributed to those activities.
The types of questions that TALIS asks are, therefore, either simple questions of existence
or type (typically yes/no or nominal multiple-choice questions), questions capturing
frequency, extent or agreement (typically count, multiple-choice or matrix questions) or
questions that are more complex in format. The latter are typically questions that combine
a relatively large number of aspects as individual items, with these sometimes combined
with simpler formats (e.g. a yes/no question combined with a question on perceived
preparedness for certain elements of initial teacher preparation). Finally, sequences or sets
of questions, for example, those that combine factual (system) information or reports of
implemented activities and attitudes towards these, can be analysed jointly to obtain a
deeper level of understanding of what exists or is implemented, of how it works or how its
impact is perceived. In general, development of TALIS questions has followed established
development principles for cross-national survey research, such as those described in work
conducted by, for example, Harkness et al. (2010[4]), Johnson et al. (2018[5]) and the Survey
Research Center (2016[6]).
Later sections of this chapter detail the process and key outcomes of TALIS 2018’s
three main development stages: the pilot in 2016, the field trial in early 2017 and the main
survey in late 2017 and early 2018 (although some data collection occurred later than this,
especially for the TALIS-PISA link and in the countries/economies that joined TALIS in
late 2018). Annex H displays the final main survey principal and teacher questionnaires
(English master version).
While this chapter describes the process and criteria for instrument development, it does
not provide in-depth details about the development of specific items or questions, nor does
it provide extensive discussion about these.
3.2. Timeline
Under a generally tight and challenging timeline, the key stages and milestones for the
instrument development were as follows (data collection phases in bold, meetings in
italics):
September 2015: First Questionnaire Expert Group (QEG) meeting (Hamburg) –
inception and agenda setting
October 2015: Content proposals and drafting of the development plan
November 2015: Nineteenth TALIS Board of Participating Countries (BPC)
meeting (Copenhagen) – review of proposals
December 2015 to January 2016: Ongoing development of the conceptual
framework and initial development work on the instruments
February 2016: Second QEG meeting (Oslo) – agreement on pilot materials
The QEG was responsible for developing the TALIS 2018 conceptual framework (see
Chapter 2) and the survey instruments. The IEA convened the group in July 2015, which
was when the TALIS 2018 Consortium began its activities. The IEA created a long list of
potential experts for the group. Some of these individuals were sourced from the IEA’s
expansive research network, some had been involved in the TALIS 2013 Instrument
Development Expert Group (IDEG) and others were individuals suggested by the BPC.
Conditional on availability and interest, the IEA, in consultation with the OECD
Secretariat, invited those experts who provided the best possible and contemporary
coverage of the topics within the TALIS 2018 purview to be members of the group. The
QEG, as eventually constituted, included the following education, policy and survey
experts:
Sigrid Blömeke, Centre for Educational Measurement (CEMO), Norway
Hilary Hollingsworth, Australian Council for Educational Research (ACER),
Australia
David Kaplan, University of Wisconsin-Madison, United States4
Daniel Muijs, University of Southampton, United Kingdom
Trude Nilsen, University of Oslo, Norway
Heather Price, Basis Policy Research and Marian University, United States5
Ronny Scherer, Centre for Educational Measurement (CEMO), Norway.
The QEG also included ex-officio members from:
The IEA: Ralph Carstens (QEG chair, consortium co-director of content),
Steffen Knoll (consortium co-director of operations), Agnes Stancel-Piatak
(analysis lead) and Deana Desa (analysis expert)
ACER: John Ainley (framework lead)
RAND Europe: Julie Belanger (liaison to the TALIS Starting Strong Survey, QEG
chair of that survey)
Statistics Canada: Jean Dumais (sampling referee)
OECD Secretariat: Karine Tremblay (project lead), Pablo Fraser,
Katarzyna Kubacka and Noémie Le Donné
Technical Advisory Group (TAG): Fons van de Vijver (TAG chair).6
The QEG began its work, the first major development phase, with a virtual meeting in
August 2015 that introduced the survey’s content focus, inputs and related information, as
well as the group’s intended work process. After that meeting, members developed a set of
draft concept notes that they reviewed at a two-day in-person meeting in Hamburg
(September 2015). Those concept notes informing the development plan were further
revised and incorporated into the conceptual framework. This process was guided by
deliberations from the BPC (later TGB) and additional input from ongoing policy dialogue
and networks among members of the OECD Directorate for Education and Skills’ senior
management team.
Having noted the general aim of reducing the number of themes (or headings) covered by
each cycle of the TALIS survey, the QEG determined that nine themes would initially
structure their work. Note, however, that the fifth theme in the following list incorporates
the two elements of teacher feedback and teacher development that were covered separately
during TALIS 2013. Note, too, that the QEG had no intention of assigning equal
importance (as expressed by, for example, response time or number of questions) to all
themes. The list below provides the themes as well as the leading QEG expert (or experts)
for each.
1. Teachers’ instructional practices and beliefs:7 Blömeke, Scherer and Nilsen
2. School leadership: Muijs
3. Teachers’ professional practices, including mobility issues advocated by the
European Commission: Hollingsworth
4. Teacher education and initial preparation: Blömeke, Scherer and Nilsen
5. Teacher feedback and development, combining teacher feedback and continuing
development: Hollingsworth
6. School climate: Price
7. Job satisfaction: Price
8. Teacher human resource issues and stakeholder relations: Muijs
9. Teacher self-efficacy: Blömeke, Scherer and Nilsen.
The BPC/TGB discussions on the content focus document resulted in the suggestion to
include questions on equity and diversity, later covered by Ainley and van den Vijver in
2016 as standalone theme number 10. The development plan also included a concept note
on the cross-cutting issue of innovation, later covered by Blömeke, Scherer and Nilsen as
a separate theme (number 11). The QEG deemed this theme to be an issue of particular
importance. Because members of the group saw equity and diversity and innovation as
themes encapsulated in the substance of each of the nine themes, they considered there was
no need (initially) to prepare separate papers on these topics. The QEG determined suitable
intersections between these two special themes and the nine general themes in order to
balance analytical potential as well as survey priorities and space.
After the initial in-person meeting, the QEG held a series of in-person or virtual meetings,
each of which lasted for about one to three hours. These occurred in December 2015
(virtual), January 2016 (virtual), February 2016 (in-person, Oslo, hosted by the Centre for
Educational Measurement), March 2016 (virtual), June 2016 (virtual), July 2016
(in-person, Hamburg, hosted by the IEA), August 2016 (virtual), November 2016 (virtual),
February 2017 (virtual), May 2017 (in-person, Paris, hosted by the OECD) and August
2017 (virtual). All meetings were scheduled in accordance with TALIS 2018’s general
development and operational timeline, that is, in keeping with the scheduled data collection
phases for the pilot, field trial and main survey, and the subsequent availability of empirical
data, feedback from NPMs and analytical outputs.
While the QEG’s core group of experts and ex-officio members remained constant
throughout the 2015 to 2017 developmental work, the IEA invited five additional academic
experts to provide specific perspectives on the pilot and field trial instruments, with their
reviews starting in May 2016 and due in June 2016. These “extended” members were:
Elsebeth Aller (Ministry of Education, Denmark; formerly Metropolitan University
College, Denmark), who provided perspectives on teaching and learning contexts
at the primary education level
Sarah Howie (University of Stellenbosch, South Africa), who provided
perspectives from low-income and middle-income countries/economies
Magdalena Mok (The Education University of Hong Kong, Hong Kong, China),
for an Asian perspective, given that no countries/economies from Asia contributed
to the TALIS 2018 pilot
Susan Seeber (Georg-August-Universität Göttingen, Germany) in relation to
teaching and learning contexts at ISCED level 3 and, in particular, vocational
education and training (VET)
Sandy Taut (Educational Quality Agency, State of Bavaria, Germany; formerly
Pontificia Universidad Católica de Chile, Chile) for a Latin American perspective.
The QEG’s later work, especially in regard to the consortium’s analysis plan, was managed
solely via desk reviews and written exchanges. As stipulated by the OECD’s terms of
reference for TALIS 2018, the QEG’s responsibilities ended with the development of the
draft main survey instruments and accompanying analytical advice and recommendations;
members were not involved in the production and/or review of the international reports.8
However, some members provided advice (Scherer and Nilsen, through the OECD
Secretariat) or were involved in reviewing draft chapters of the report in early 2019 (Price,
through the consortium).
3.4.1. Proposing new, revised and retained materials for ISCED level 2
The TALIS 2018 consortium, in close co-operation with the QEG and the OECD
Secretariat, and seeking advice from the TAG at key stages, moved from the development
plan to the second major phase of developing the teacher and principal questionnaires.
During this second phase, the consortium generally addressed all interests and needs voiced
by stakeholders or indicated by the literature cited in the development plan.
The consortium presented the first drafts of the instruments at the TGB’s first meeting in
Singapore in March 2016. Subsequent feedback included written comments from the board,
additional reviews from the extended QEG, additional feedback from institutional partners
and observers to the TGB (such as Education International), as well as from the OECD
Directorate for Education and Skills’ senior management. The work at this stage also
focused on earlier input from policy work associated with the TALIS Initial Teacher
Preparation (ITP) study and from initial work on the TALIS Starting Strong Survey.
During the meeting in Singapore, the TGB considered and then provided advice on the
extent to which the content of the survey instruments needed to be balanced between
existing questions from TALIS 2013, questions designed to improve the measurement of
existing constructs and questions addressing new topics that had emerged within the
nominated themes. To yield the type of feedback that would advance the instrument design
parameters from the initial content focus and incorporate the consortium’s development
proposals, meeting participants worked in small groups and developed plenary summaries.
Suggestions on improving the measurement of existing constructs were generally based on
reflections on analyses of TALIS 2013 data. Introduction of questions reflecting new topics
within the survey themes typically arose from scrutiny of recent research literature or from
areas of interest identified by countries and economies or by institutional stakeholders.
Increasing synergy between TALIS and, most importantly, the TALIS Starting Strong
Survey and the PISA 2018 survey contexts, yielded other areas of interest. It soon became
clear that effort to fully implement all desired changes and additions could result in the
instruments not retaining a sufficient amount of (key) content that would elicit the data
needed to support analyses of changes over time. The TGB was also mindful of the
necessity to review any proposed extensions (and their intended indicators) to the
instruments in terms of survey time and burden for teachers.
In pragmatic terms, the TGB wanted to achieve a balance of instrument content within an
average response time (for the English version) of 45 minutes. However, the board also
recognised that the issue of time itself was still under review and would not be resolved
until after the field trial when reliable timing information based on larger volumes of data
would be available. In the meantime, the TGB invited the OECD Secretariat and the TALIS
2018 consortium to use evidence from previous rounds of TALIS as the basis of discussion
on questionnaire length and the potential consequences of varying questionnaire lengths
for data quality. The consortium accordingly re-analysed timing and other process data
from the 2013 field trial in terms of survey fatigue and disproportionate increases of item
non-response as a function of time. The 2013 data indicated that while the percentages of
respondents who did not complete the survey varied substantially across
countries/economies and contexts, the percentages generally increased in a linear way
along the question sequence. The actual time it took respondents to complete the survey
also varied considerably, with the averages exceeding the 45-minute target in many cases.
Limitations with respect to time (including those for national additions) also directly
affected ability to revise materials, introduce new materials or keep core questions constant.
Because decisions on these matters would be necessary and feasible only when the main
survey administration began towards the end of 2017, the consortium foresaw the need to
conduct a priority rating exercise that would facilitate work directed towards managing and
reducing survey content after completion of the field trial.
The consortium presented NPMs with drafts of the instruments during the first NPM
meeting in April 2016. The NPMs, in turn, provided helpful feedback relating to local
relevance, translatability and validity in general. The consortium collated feedback
pertaining to survey implementation, including clarity of terminology, and shared and
discussed it with the QEG. As had occurred during previous development rounds, deep
discussion emphasised instrument length and related response burden and resulted in a
recommendation to significantly reduce both before the field trial and then again prior to
the main survey.
More specifically, the review of the pilot questionnaires focused on the following key
dimensions:
applicability of concepts and validity
level of complexity of the questionnaires
organisation of topics and items
applicability of items across ISCED levels and programmes (academic/vocational)
international versus local applicability of items
item wording and definition of terms
appropriateness and cultural relevance
mandatory national adaptations
foreseen translation issues
flow of questions (overall and specifically with respect to skipping instructions)
length of questionnaires.
The TALIS 2018 Consortium planned that around ten countries/economies would
contribute to the pilot by translating the generic source version of the questionnaires from
English into the national target language and by having a group of target population
members review the questionnaire items under the oversight of the respective NPMs.
However, when the schedule for this work proved to be too tight, the possible solution
offered was that of making only the mandatory national adaptations to the generic English
source version, provided that all focus group members were bilingual. Eventually, 11 of
these countries and economies contributed to the pilot study.9 It is important to note that
there was never an intention to test applicability and functioning of the pilot questionnaires
in all national settings and contexts, that is, in the 48 countries, economies and educational
systems that participated in TALIS 2018.
The pilot instruments (along with a draft glossary) were finalised in late April 2016 and
released for the data collection in May 2016. The consortium held preparation webinars
with representatives from the pilot countries/economies in late March and early April 2016
and also released a comprehensive guidance manual that included a concise list of general
and per-item probing questions. The pilot was conducted in May 2016.
Six countries/economies translated the questionnaires into one national language and one
country translated them into two languages. The remaining four countries/economies used
the English source version of the questionnaires with necessary local adaptations. The pilot
was conducted using paper instruments and the only data entry work required was that
NPMs filled in an online structured-session feedback questionnaire at the end of the focus
group work. To manage the response burden at the pilot stage, the consortium and QEG
split the teacher questionnaire into two partially overlapping versions, A and B. Although
the QEG expected each focus group to discuss only one part in detail, members of each
group received the full teacher questionnaire so that they could see and understand the full
scope of topics, materials and contexts. There was no commensurate split for the principal
questionnaire.
The pilot countries/economies had used various convenience (i.e. non-probabilistic)
methods to gain teacher and principal collaboration (e.g. support from the ministry of
education; using existing ties with teachers’ networks developed during earlier research
projects). The pilot countries/economies also reported using a variety of incentives and
rewards (i.e. a combination of monetary and non-monetary support as well as giveaways)
or none at all. Most of the principal and teacher focus groups consisted of about five to
ten people. In general, countries/economies tried to maintain a balance between male and
female respondents, private and public schools, and urban and rural representatives, as well
as between age groups. However, few participating principals were younger than 50 years
old. Total average session duration for answering and discussing one instrument in the
focus group was about three to three-and-a-half hours.
In terms of response burden, the majority of principals reported taking, on average, about
60 minutes to respond to the principal questionnaire, with individual times ranging from
about 30 to 80 minutes. Because the teacher questionnaire was administered in two versions
(A, B), each of the two groups of teachers responded to and discussed about two-thirds of
the materials. Some teachers were able to complete the questionnaire (again, about two-
thirds of all teacher questionnaire materials) in as little as 30 minutes, whereas some
required up to 2 hours. The average time for completing the partial teacher questionnaire
was similar to the principal questionnaire, that is, 60 minutes, indicating that teachers would
need 90 to 95 minutes to complete the full set of teacher questionnaire materials.
In general, the results and inputs from the focus groups reflected various aspects relevant
to development of the questionnaires. These included the acceptability and relevance of the
questions, the clarity of the language used in the questions, ambiguity with respect to
terminology, overburdening in terms of response detail or the need to recall past events and
facts, and preferences for the alternatives presented to respondents (e.g. two versions of
career plan questions or regarding preferred resource allocation). While most of the
feedback was incidental (i.e. raised in just 1 of the 11 countries), some of it exhibited
commonality and some of it clearly represented opposing views and feedback, with some
of the latter impossible to reconcile. The results also showed that teachers and principals
did not indicate any substantial areas, topics or issues that they might have thought were
overlooked during the TALIS 2018 thematic scoping. As expected, respondents repeatedly
reported that the questionnaires took too much time to complete.
The input and recommendations from the extended QEG, collected in June 2016, were
equally rich in nature, such as those relating to the applicability of conceptual deliberations,
literature and earlier insights from research in the context of low-income countries or in
particular regions. When the experts were asked to reflect on the applicability of universal
questionnaire templates for ISCED levels 1 and 3 (including VET), they overwhelmingly
confirmed that the materials were relevant and applicable at all targeted levels of education.
However, they did recommend the inclusion of particular materials that were not only
relevant at ISCED level 1 but also aligned with TGB interests.
initially agreed-on deletions, changes and additions were compiled in time for
consideration at the second QEG meeting of the TALIS Starting Strong Survey. During
this meeting, which took place later in July 2016, the experts from the early childhood
perspective noted and reflected on the changes that the TALIS QEG had made to the
materials that needed to be aligned between the two surveys. The early childhood experts
also made several salient recommendations for changes to the TALIS materials in terms of
language or clarity, as well as changes that they thought would improve the conceptual link
and analytical potential between ISCED levels 0 and 1.
The TALIS QEG members subsequently incorporated a relatively large number of edits
and changes into the TALIS materials to improve the clarity of the language used (i.e. at
the word level). Editing also focused on harmonising language across questions and items,
across response options and between the two surveys (TALIS and TALIS Starting Strong),
with the latter involving close collaboration between both QEGs. Examples of the latter
kind included, but were not limited to, consistent question instructions; consistent response
options; the highlighting of required adaptations; the consistent use of certain terminology
(e.g. “migration background”, “children and young people”, “parents or guardians”); and,
finally, the use of the personal pronouns “your/my”, but only in those instances where a
proximate personal characteristic of the respondent was meant (in contrast to, say, “your
school”). Such changes were also applied to questions retained from TALIS 2013, provided
that the QEG was assured, after careful consideration, that the change would not have an
impact on the statistical properties of the responses. Some of these changes were
nonetheless recommended for experimental comparison as part of the field trial, in the
interest of retaining the revised version or of reverting to the original in the main survey.
Some of the key substantial and structural revisions to the principal and teacher
questionnaires were the following:
The merging of two sets of questions relating to initial teacher preparation: The
first set concerned elements covered in initial teacher preparation and asked the
teachers to give a self-reported rating of their level of preparedness for each
element. The second asked teachers to identify, from a listing of subjects, if they
had specialised in any of them during their initial teacher preparation and whether
they were currently teaching the subject. The intention behind both mergers was to
extend analytical possibilities, with the latter also relating to a corresponding
question in the PISA 2018 teacher questionnaire.
A re-allocation of a set of questions relating to self-reported activities and practices
(conceptually related to instructional quality dimensions): These received separate
sets of response options depending on whether or not the practices were directly
observable (low/high or shallow/deep inference).
Changing the format of some of the questions that allowed multiple responses:
These were extended, to the extent possible, to a yes/no format. Some questions
were similarly converted from detailed counts (e.g. for certain categories of staff)
to censored intervals (e.g. “1–5”, “6–10”, etc.) on the premise that these are easier
to respond to.
Revisions to materials relating to the theme of equity and diversity: These changes
were the result of changes in items related to the global competency theme in PISA
2018, and of the TGB and NPM feedback received in March and April 2016. In
particular, filter questions were used in cases where practices related to equity and
diversity did not apply according to principals and teachers because of the local
and indicators allocated to each area and theme. The QEG assessed, as a crude and,
arguably, imperfect measure, that the correlation between these 2 metrics (priority and
volume of material in terms of data points, questions and pages) was approximately equal
to 0.8 for both the principal and teacher questionnaires. While this correlation reflected a
general match between the development work at that stage and the TGB’s mandate, the
match was not perfect in all areas, in particular for school leadership in the principal
questionnaire and for the combined theme of feedback and development in the teacher
questionnaire.
This situation presented the QEG with an opportunity to choose the most appropriate
balance of materials for the field trial and, by extension, the main survey. This opportunity
aligned with the stipulation that, in keeping with technical standards, statistical
considerations, and established practices in large-scale assessments, no new materials
could be developed or admitted to the survey after the field trial.
The main concern in regard to the teacher questionnaire was that the volume of all material
in it meant a response time of about 90 minutes in total. However, the total amount of all
material that the field trial could accommodate under a form-based design meant a response
time of 75 minutes, including a maximum time of 5 minutes to answer the national
additions. As a result, the QEG needed to remove about 20% of the material from the field
trial teacher questionnaire.
Alternative versions (forms) of the teacher questionnaire were used in a rotated design for
the field trial. This approach meant the total amount of material could be trialled while
ensuring that the response burden for any one teacher would not exceed the targeted
response time for the main survey instrument, that is, 45 to 60 minutes on average for the
English version, unless analysis of timing, fatigue and non-response indicators suggested
that the total response time of 45 minutes should be closer to the time stated originally in
the survey’s terms of reference.
A second reason for using the rotated design related to the need to trial some alternative
question wordings or formats, while a third reason came from the TAG’s recommendation
that the relative position of at least one section of the teacher questionnaire should alternate
between forms to allow study of positional effects related to, for example, fatigue, effort,
or other order effects, such as priming. Full randomisation of sections was not feasible
because TALIS needed to continue to rely on a logical order of survey themes, from initial
preparation through to current teaching activities and overall satisfaction with teaching.
The QEG and consortium rejected suggestions to implement more complex designs, that
is, those with more than three different forms, for reasons of operational complexity,
especially in regard to paper-based administration of the survey instruments and in regard
to analytical requirements.
To collect the views and inputs of TGB members on what materials to retain/delete, the
consortium, in consultation with the OECD Secretariat, developed an Excel format input
sheet for the teacher questionnaire (only). The sheet listed all developed questions, the key
links to sibling activities (PISA and the Starting Strong Survey), an estimate of response
time by question and section, and the QEG’s identification of possible reductions. Because
questions deemed important for cross-cycle comparison were locked in, they were not
eligible for deletion. The consortium asked the TGB members to indicate, for each
question, whether they thought it should be dropped from the field trial teacher
questionnaire. The consortium also asked members to keep in mind, when making their
decisions, the target response time of 75 minutes total.
The following tables illustrate the final field trial approach, which combined the TGB’s
ratings and wider considerations. Table 3.1 lists the questionnaire sections used in the field
trial along with the number of core questions determined after the TGB meeting. It also
gives the number of co-ordinated national options (CNOs) – questions that were not part
of the international core questionnaire content but ones that several countries/economies
wanted to leave in the questionnaire templates (note that CNOs counted towards the
allowance for national adaptations; see Chapter 4 for details).
1. Consistent with TALIS 2008 and 2013, a section on classroom-based characteristics, activities and
perceptions was contextualised to a specific “target class” (i.e. group of students), operationalised as the first
class that teachers taught in the school after 11 a.m. on the previous Tuesday.
Table 3.2 lists the sequence of blocks included in each of the three questionnaire forms.
The design followed substantial considerations more than statistical ones. The inclusion of
Section A on background in each version enabled basic breakdowns of the data during the
field trial analyses. Each of the other two sections had two forms that differed in number
of sections. However, the two sections balanced overall with respect to the number of
questions and the estimated amount of time needed to respond to materials (between 58
and 59 minutes per questionnaire, including national options of up to 5 minutes). The page
count (about 24 to 25 pages) was roughly equivalent between versions but varied by
national version and the degree of text expansion following translation.
The asterisks (*) in Table 3.2 indicate sections that included an experimental version. The
experimental version was one that the QEG included to assess the functioning of a revised
approach to measuring an indicator and one that the group expected would not be
immediately successful and therefore would not be retained in the main survey.
Intentionally, Section G on school climate and job satisfaction appeared in two sequences
designed to assess primacy and recency effects: once in a late final position, as was the case
in TALIS 2013 (Form B), and once in an earlier position (Form C). Section H was included
as the last section because it was fairly experimental at the field trial stage and its continued
inclusion had yet to be formally decided.
x : not applicable
The consortium considered the size of the sample required for the field trial in light of this
instrument structure (see also Chapter 5). Because a slightly larger number of schools and
teachers were needed for the field trial in 2018 than in 2013, the consortium fixed the
number at 30 schools and 20 teachers within each sampled school. Attrition aside, either
600 (common sections), 400 (materials used in 2 of 3 forms), or, as a minimum, 200 data
points (for question alternatives/experiments) would then be available per
country/economy for analyses. The above design made it possible to analyse all themes
with one another even though this possibility was not a key design goal.
Given the much smaller size of the school/principal population (about 30 schools per
country/economy), the use of a rotated administration design for the principal questionnaire
was not feasible. Timelines, resource constraints and planned analyses also precluded a
rotational or otherwise partial design in the main survey, which meant only a
single/common teacher questionnaire could be used for each target population.
QEG was confident, after careful consideration, that these questions would not affect the
statistical properties of the responses.
Review of experiments
The following bulleted content describes how the experiments were embedded in the field
trial and gives an account of the main insights gained from their presence:
Alternative frequency scales for “extent” questions: Some of the existing questions
from TALIS 2013 used an established set of response options (“not at all”, “to some
extent”, “quite a bit”, “a lot”), a format also used in PISA and, in part, taken from
earlier cycles of TALIS. The QEG for TALIS 2018 raised concerns with respect to
the exactitude of the “quite a bit” option and challenged the appropriateness of the
response options. As an alternative, the consortium proposed this set of response
options: “not at all”, “very little”, “to some extent”, “a lot”. The experiment was
implemented in the established (i.e. trend) self-efficacy question, with the original
set comprising one form and the revised set the other form. Descriptive statistics
were used to evaluate each alternative, while measurement invariance analyses
along with cross-country/economy, cross-population and cross-cycle scale
validation were used to evaluate the scale reliability and validity of the two
alternatives. This work led to the established 2013 version being retained despite
some differences in the measurement properties and the response frequencies.
Alternative wording for family-related aspects: All instances of the word “family”
were changed to the words “personal life” in the 2018 questionnaires. The change
acknowledged that not all teachers have family duties and ties. The experiment,
therefore, included “family” in one form of an item on barriers to professional
development and “personal life” in the other form. This approach made it possible
to check the equivalence of response proportions and other characteristics
(e.g. non-response). Descriptive statistics were used to evaluate the item properties
of the two scales (i.e. one with “family” and the other with “personal life”).
Measurement invariance evaluation along with cross-country/economy, cross-
population and cross-cycle scale validation were used to evaluate the reliability and
validity of each scale. The established item formulation from 2013 was
subsequently retained given evidence of some substantial between-
country/economy changes in the measurement properties.
Substantial revisions of a question on teacher collaboration: The aim of this
experiment was to contrast the original (2013) and revised version of the question
for statistical equivalence, with the expectation that one of the two versions would
be retained, depending of course on the outcome of the evaluation. However, the
members of the QEG noted a preference for retaining the revised version because
they considered it had greater applicability to TALIS 2018. Descriptive statistics
were used to evaluate the item properties of the two scales, while measurement
invariance evaluation along with cross-country/economy, cross-population and
cross-cycle scale validation were used to evaluate the reliability and validity of the
two scales. As before, the 2013 formulations of the items were retained because of
substantial changes in the measurement properties.
Context priming in question stems: The intention behind these experiments was to
investigate the impact of priming respondents by including context-related
information in the question stem versus not including such information
(i.e. a context-free stem). The question used for this purpose was a school climate
question from TALIS 2013. The experiment used that version of the question as
well as an alternative version that included a stem that read: “Thinking about the
general climate in this school, …”. Once again, descriptive statistics were used to
evaluate the item properties of these two alternatives and measurement invariance
evaluation along with cross-country/economy, cross-population and cross-cycle
scale validation were used to assess scale reliability and validity. As before, the
2013 stem was retained because of substantial changes in the measurement
properties.
time plus 5 minutes to respond to national additions was somewhat met but only on
average across the participating countries/economies.
The average gross response time for principals of about 76 minutes was somewhat
higher across the ISCED level 2 countries and economies.
The average net time (across countries/economies and respondents) for the
principal questionnaire was around 75 minutes for the English-speaking
countries/economies and 85 minutes across all countries/economies, with the range
of time spent extending from about 1 to 2 hours. Thus, the time targets expressed
in the questionnaire’s prologue (i.e. 45 to 60 minutes) were, on average, exceeded
during the field trial.
A high level of consistency was evident for both the net and gross times across the
ISCED levels and the TALIS-PISA link within countries/economies. Thus, for
example, average gross and average net times for ISCED levels 1 and 2 were
largely consistent for both teachers and principals. The QEG members used the
averages for the ISCED level 2 countries/economies as the benchmarks for their
work.
However, as occurred during TALIS 2013, the average net and gross times varied
considerably across the TALIS 2018 countries and economies.
A high level of correlation was evident in the average gross times taken to complete
the teacher and principal questionnaires administered in the same language. The
average per-question time for English-speaking versions and all languages
correlated at 0.99. However, the totals for the English and the non-English versions
were very different, with the averages for all languages consistently higher. This
finding meant that the time target for the English version, as stipulated by the
survey’s terms of reference, would not be a meaningful or fair estimate for response
burden across such a variety of countries/economies and contexts. The QEG
members, therefore, used the empirical times across all languages in the ISCED
level 2 survey for their deliberations.
Stark differences were identified between the initial crude estimates of response
time (e.g. those used at the second TGB meeting) and the actual empirical times.
The QEG, therefore, used the empirical times for ISCED level 2.
The time respondents took to read the questions versus the time they took to
respond to them varied. Generally, questions with fewer data points and less
information collected required more time to read (on average 60% of the time spent
on the question) than time to respond (around 40% of the time spent on the
question). Questions with more data points, mostly longer matrix-type questions,
required relatively less time to read (about 25% to 30% of the time) than time to
respond (around 70% to 75% of the time). The total time taken for a question
correlated highly with the amount of information it collected. However, some of
the outliers present in questions entailed a more complex and longer recall/response
process (e.g. total time worked at school in last complete week; staff counts by
category). When considering the amount of time respondents across all cases
(e.g. ISCED levels) would need to answer the questions remaining after the field
trial, the consortium estimated that 25% of the total time would be spent on reading
and 75% of the time on responding.
Technical validity: All items had to have proven measurement quality, meaning
that no new constructs, questions or items could be considered. If field trial data
indicated a fundamental defect, the construct, question or item was dropped rather
than revised. Minor edits relating to grammar, spelling, capitalisation and
consistency were accepted. However, no substantive or semantic changes were
considered unless sufficiently indicated by field trial data, for example, by
responses to open-ended “Other, please specify” questions.11
Conceptual value: The QEG considered the conceptual value of items in relation
to the TALIS 2018 conceptual framework and the OECD’s draft reporting plan.
Members also took into consideration linkages to the TALIS Starting Strong
Survey, the TALIS Video Study and PISA 2018, as well as possible contributions
to the Sustainable Development Goals (SDGs) Target 4 and, as a priority
consideration, to TALIS 2013.
Efficiency/length: Analysis of the timing data from the field trial signified that the
overall length of both questionnaires had to be reduced by at least one‑third. The
meeting sought, wherever possible, to reduce the number of items per question
while ensuring the most efficient coverage of constructs in the instruments. Short
questions (in terms of reading or response times) were preferred to longer ones, as
were questions that served the needs of multiple themes. Where two or more
questions covered similar constructs, all but one was dropped, with the decision
based on the need and priority to report changes over time, criterion validity or
similar rationale. In those instances where a construct was assessed from
two different perspectives, the perspective with the better psychometric quality was
the one retained.
In addition to removing almost all of the open-ended “please specify” dimensions, the
meeting members deleted around 32 items from 15 questions in the principal questionnaire,
and 65 items from 21 questions in the teacher questionnaire. In three instances, the QEG
collapsed two or more dimensions into a single dimension.
Key modifications made to the structure or scope of questions were documented, annotated
and communicated to the TGB, while minor modifications to spelling, grammar,
punctuation and wording were made in accordance with the agreed approach and rules.
Examples of deletions made on the basis of scientific rationale (poor functioning, low
comparability) or other rationales (e.g. similar information or proxies available elsewhere)
included the following: all situational judgement12 items on cognitive activation, classroom
management, clarity of instruction and teacher support (cross-cultural variation very high,
bad model fit overall); job commitment/career plans (attrition proxy could be derived from
teachers’ age and number of years remaining in teaching); and a school team’s degree of
innovativeness (same scale administered to teachers and principals, teachers able to provide
more accurate ratings).
Given increased ethical concerns about the collection and use of data, including process
data, the consortium extended the questionnaire introductions to include language that
informed respondents that process data, such as time spent on items, might be collected in
an anonymous way and used for methodological and validation purposes.13
After eliminating questions based on this scientific rationale, the QEG considered
three additional categories of possible deletions based on their understanding of priorities,
on the analytical potential of questions and on ensuring that balance breadth and depth
would still be maintained for each of the themes. Accordingly, the QEG flagged a sufficient
Table 3.3. Overview of question count across TALIS 2008, 2013 and 2018 (field trial and
main survey)
2008 main survey 2013 main survey 2018 field trial 2018 main survey
Principal 37 questions 39 questions 55 questions 45 questions
questionnaire No CNOs No CNOs No CNOs
Increase (core):
6 from 2013
These estimates meant that the gross average response time stipulated by the survey’s terms
of reference, that is, 45 minutes for the English version(s), would be exceeded. Given the
consortium’s concern about the overall increase in the number of questions, its members
presented key findings from the process data analysis and the preliminary response time
estimates to the TAG during the June 2017 meeting. In response, the TAG advised that it
did not see a strong technical reason for not using the proposed questionnaire materials or
for a specific cut-off time. However, the TAG did express concern that a longer
questionnaire would likely yield such disadvantages as a hastier response style and poorer
quality responses. The TAG members, therefore, suggested that process/time data be
included in the data editing and the review of “straight-lining” (including defining item-
level non-response) of the main survey data. Because of time constraints, priorities and
overall limited resources, it was not possible to use the process data in this way or to analyse
the sample process data (collected as part of the main survey) by the time this chapter was
completed.
The consortium concluded that TALIS 2018 reached a limit, firstly with respect to what
the current survey and questionnaire design could accommodate in terms of breadth and
depth within and across the sizable number of themes, and second with respect to what it
could reasonably expect from principals and teachers in terms of survey engagement and
response burden. The consortium advised the TGB to expect, in comparison to TALIS
2013, the following: (1) some reduced co-operation and thus a higher level of questionnaire
non-response; (2) some increase in the level of non-response for questions towards the end
of the questionnaire; and (3) some reduction in response quality. While Aspects 2 and 3
had not been fully analysed at the time of writing, there was initially no evidence to indicate
a noticeable effect on response rates across countries (Aspect 1) and cycles, as was reported
to the TGB during its fifth meeting in Paris in February 2019.
one answer from its substantive part (i.e. non-background), discussions with the TALIS
Board of Participating Countries (BPC) resulted in the board’s decision to retain the
threshold at one answer and its agreement that this criterion would, therefore, suffice to
define a “participating teacher” (Technical Standard 4.28 in TALIS 2013).
During discussions between the TALIS 2018 consortium and the TAG, it was suggested,
via an internal technical memo, that the notion of “participating teacher” should be
investigated again and that the impact of various rules should be assessed through the use
of TALIS 2013 teacher data. The data set used was one where the records contributed to
most, if not all, of the TALIS 2013 tables and models, and it was therefore seen as
preferable to a set where useful information was scarce.
To achieve the analysis, the consortium compiled a list of some 46 variables and 3 scales
comprised of 70 individual items in total (List 1) from the teacher background and initial
preparation section of the TALIS 2013 teacher questionnaire (23 items), the professional
development scale (23 items), the self-efficacy scale (12 items), the professional
collaboration scale (4 items) and the job satisfaction scale (8 items). Meanwhile, the OECD
Secretariat proposed a second list (List 2) consisting of 12 items from the teacher
background and initial preparation section of the questionnaire (3 items) and from the
professional development section of the questionnaire (9 items). List 1 spanned the entire
teacher questionnaire and addressed key TALIS 2013 topics (which remained important in
TALIS 2018). List 2 was optimised to address UN Sustainable Development Goal (SDG)
4.
The consortium then tested 3 different definitions of a “participating teacher”: (1) where
the teacher’s responses covered at least 75% of the 46 variables and 3 scales on List 1;
(2) where the responses covered at least 50% of the 46 variables and 3 scales on List 1; and
(3) where the responses covered each of the 12 variables in List 2. A response to or no
response to any other item or variable was not considered a factor in these definitions.
For each definition, the consortium assessed and classified each returned TALIS 2013
teacher questionnaire as “participating teacher” or “non-participating teacher”. Teacher
non-response adjustment factors were recomputed if needed. If the number of
“participating teachers” fell under 50% of the selected teachers, the consortium flagged the
school as “non-participating” and then recomputed the school non-response adjustment.
Final estimation weights and replication weights were also recomputed.
To assess the impact of each definition, the consortium recomputed estimates and
compared them with the results of the TALIS 2013 cycle for a number of tables. The
variables covered during this process included: (1) age and gender distributions;
(2) teachers’ educational attainment; (3) hours spent working and spent teaching;
(4) teachers’ practices and job satisfaction; (5) subjects taught in current year;
(6) professional development activities; and (8) professional development needs. The
impact on participation rates and, therefore, on adjudication recommendations was also
looked into.
In summary, and probably because of the generally high response rate in TALIS, none of
the definitions had a severe impact on the estimates: the differences between the estimates
published in 2013 and the alternative versions studied were for the most part negligible.
Definition 2 (50% of items on List 1) was the most lenient and had the least impact on the
estimates and on adjudication as hypothesised. Definition 1 (75% of List 1) and
Definition 3 (all of List 2) had more impact than Definition 2. The impact also differed
across countries/economies because both lists spanned the teacher questionnaire to
different extents. Definition 1 tended to have the most impact on teachers who became
fatigued early on while answering the questionnaire. However, because of the shorter span
of List 2 and because the items on it appeared mostly at the beginning of the teacher
questionnaire, the impact of early drop-outs was not as clearly felt as it was under
Definition 3.
All 3 definitions had little impact on those countries/economies where teacher and school
participation rates were well above 85%. The impact was much more visible in
countries/economies where participation was moderate (say, from 60% to 75%) or where
respondents experienced survey fatigue relatively early while answering the
questionnaires. Adjudication recommendations stayed identical except for two countries
that would have seen their rating drop one category (from “good” to “fair” and from “fair”
to “poor” respectively) due to loss of “responding teachers” and the ensuing loss of
“participating schools”.
The detailed results of this study were presented to the TAG in June 2017 at their meeting
in Paris. The TAG recommended that TALIS 2018 Technical Standard 3.29 should be
modified to reflect Definition 1 (after the corresponding teacher questionnaire items in
TALIS 2018 had been identified)14. The TAG further recommended that data records which
did not meet the standard on teacher participation should be: (1) excluded from the
respective analyses; or (2) included, but with a note advising that the participating teacher
standard had not been met.
During its third meeting in Lisbon in July 2017, the TGB decided not to adopt the definition
but to apply the existing Technical Standard 3.29 instead (see Chapters 8, 9 and 10 for more
details).
References
Harkness, J. et al. (eds.) (2010), Survey Methods in Multinational, Multiregional, and [4]
Multicultural Contexts, Wiley Series in Survey Methodology, John Wiley & Sons, Hoboken,
New Jersey.
Jensen, B. and S. Couper (2015), “Joint conceptual framework for TALIS and PISA Synergies”. [3]
Johnson, T. et al. (eds.) (2018), Advances in Comparative Survey Methods: Multinational, [5]
Multiregional, and Multicultural Contexts (3MC), John Wiley & Sons, Hoboken, New Jersey.
OECD (2015), “Development Plan for the OECD Teaching and Learning International Survey [2]
(TALIS) 2018”, EDU/INES/TALIS(2015)7 (internal document), Directorate for Education
and Skills, OECD, Paris.
OECD (2015), “Guiding the Policy and Content Focus of TALIS 2018”, [1]
EDU/INES/TALIS(2015)3 (internal document), Directorate for Education and Skills, OECD,
Paris.
Survey Research Center (2016), Guidelines for Best Practice in Cross-Cultural Surveys, Institute [6]
for Social Research, University of Michigan, Ann Arbor, MI, http://ccsg.isr.umich.edu/
(accessed on 27 March 2019).
Notes
1
At this stage, it was determined that the concurrent uptake of the TALIS-PISA link option as part
of both TALIS and the PISA teacher questionnaire would constitute a significant overburdening of
teachers, especially in small countries where a large proportion of schools would be participating in
both TALIS and PISA, and thus could mean some teachers receiving up to three different teacher
questionnaires (core TALIS, TALIS-PISA link and PISA).
2
The TALIS Board of Participating Countries (BPC) became the TALIS Governing Board (TGB)
from 1 January 2016.
3
The TAG also provided input through a series of virtual meetings (each lasting around two to three
hours) conducted throughout the duration of the study: January 2016, March 2016, June 2016,
November 2016 and March 2018. Written consultations about specific issues also took place during
this time period.
4
Dr Kaplan’s contribution to the QEG was methodological in nature rather than related to the
survey’s core and substantial themes. Dr Kaplan also provided an important point of liaison with the
PISA 2015 and PISA 2018 QEGs as required by the TALIS 2018 terms of reference.
5
Dr Price provided essential links to the analytical work conducted as part of TALIS 2013, and
which she contributed to at the time.
6
Dr van de Vijver later contributed to the substantial development of the survey in connection with
the topic of equity and diversity. This topic was changed into a full theme rather than a cross-cutting
aspect prior to the field trial.
7
Eventually, the TALIS 2018 main survey instruments did not cover “beliefs” about teaching, given
sub-par measurement characteristics in the field trial (and originally in TALIS 2013). The theme
title was subsequently shortened in the final conceptual framework.
8
As agreed with the OECD Secretariat and TGB, QEG members received advance access to the
draft international database in mid-2019 so that they could conduct their own analyses and
publications, with that access conditional on the same embargo and data use policy that applied to
the TGB.
9
Pilot contributors were Alberta (Canada), Australia, Colombia, the Czech Republic, Dubai (United
Arab Emirates), France, Kazakhstan, the Netherlands, Norway, Portugal and Sweden.
10
Two late-joining countries/economies out of the total 48 countries/economies conducted a field
trial at a later stage; the results did not contribute to the main field trial analyses.
11
The European Commission requested the inclusion of an indicator on the proportion of students
with an immigrant background (first or second generation) in the school (principal questionnaire)
and classroom (teacher questionnaire). This late addition was problematic from a process
perspective because the question had not been field-trialled. Given the European Commission’s
strong interest in conducting analyses of school and teacher data in light of the concentration of
immigrant students, the OECD Secretariat and TALIS 2018 Consortium decided to defer the
decision to the TGB. The TGB subsequently decided, during its third meeting in Lisbon, to include
this question in the main survey.
12
A variety of approaches were used to analyse these items in co-operation with Dr Leslie
Rutkowski and in consultation with the TAG as well as additional experts convened by the OECD
Secretariat for a webinar. Most importantly, insufficient measurement properties (in item response
theory and factor analytical frameworks) and the absence of a consensus on appropriate or preferred
situational reaction led to the QEG dropping the items. The TAG recommended that future TALIS
cycles explore different scale scoring methods, such as “wisdom of the crowd” or expert ratings.
13
For the main survey, the consortium collected process data (i.e. timestamped events reflecting
user actions, browser and screen size details, but no IP addresses) from a sample of respondents
only. The intention was to use the data to assess the overall quality and validity of the data, for
example, with respect to proportions of respondents with exceptionally low or high response times.
14
The consortium and QEG determined the following list of 20 key questions from the field trial
that could reasonably describe a teacher, that had high policy relevance and that were assumed to be
relatively stable across cycles: gender (TQ-01), age (TQ-02), highest educational attainment (TQ-
04), qualification elements and preparedness (TQ-08), employment status tenure (TQ-11),
employment status full-time equivalent (TQ-12), work experience (TQ-13), special needs teaching
status (TQ-16), subjects taught (TQ-17), time distribution – total hours (TQ-18), time distribution –
teaching hours (TQ-19), professional development types/formats (TQ-27), professional
development topics (TQ-28), professional development needs (TQ-32), engagement in collaborative
activities (scale, TQ-40), general self-efficacy (scale, TQ-41), satisfaction with classroom autonomy
(scale, TQ-52), core teaching practices in target class (scale, TQ-54), satisfaction with the profession
and school (scale, TQ-65), and self-efficacy in multicultural environments (scale, TQ-70).
This chapter summarises the outcomes and procedures for national adaptation verification
and international translation verification, and includes information about the layout
verification process of the Teaching and Learning International Survey (TALIS)
instruments. The international adaptation, translation and layout verification of all
specified instruments in the identified target languages was conducted before the field trial
and main data collection of TALIS 2018.
All participating countries/economies produced national survey instruments efficiently and
of high quality. Quality control during production of the national instruments helped
ensure the collection of high-quality data that can be compared internationally across
countries/economies and over time. National adaptation forms supported adaptations of
the international source versions in the main survey for all target populations chosen by
the participating countries and economies. The materials for each country/economy and
survey language were carefully cross-checked against the various source instruments, that
is, the international (English, French) instruments, the versions for the different target
populations and the versions used in the previous TALIS cycle (TALIS 2013).
The development of the various target questionnaires produced from the international
source questionnaire, ranging from one to four different target populations (if and when
applicable), was generally successful with respect to ensuring accuracy and consistency
not only within the national materials for each country/economy but also across the
national survey instruments used by each participating country/economy.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
The TALIS 2018 survey instruments were reviewed during several stages of their
preparation. Instruments included questionnaires (online and paper data collection) and
cover letters (online data collection). The reviews focused on approving national
modifications to all instruments (in a given target language) and this chapter describes the
procedures related to that work, which encompassed three major activities:
translation and adaptation of the international source versions of the TALIS
instruments into the national languages
international verification of the national translations and adaptations
international layout verification of the final national instruments.
Everybody involved in preparing the instruments had to meet the procedural requirements
for translation and translation verification (including the submission and review
procedures) outlined in the TALIS 2018 Technical Standards 4.1–4.27. The TALIS 2018
Survey Operations Procedures Unit 3, provided to all TALIS 2018 national project
managers (NPMs), contained detailed instructions on instrument preparation.
The survey instruments, which included an international version in English and a version
translated into French (the official languages of the OECD), were released to the national
study teams during three key phases. The first phase consisted of a pilot study, during which
experts reviewed the survey instruments. Each national study centre could use either the
English or French source version of the instruments (on the assumption that the teachers
and principals responding to them could understand English or French) or perform a full
translation into the local language(s). Because the piloted national TALIS questionnaires
were used solely to collect qualitative data and feedback rather than quantitative,
internationally comparable data, national adaptations and translations were not examined
by external verifiers.
The second and third key phases were the field trial and the main survey respectively.
During these phases, the national questionnaires underwent stringent independent
adaptation, translation and layout verification processes. The national study centres began
this work by adapting the international source version of the teacher and principal
questionnaires (available in English and French) to fit their respective national or
subnational contexts. This work included structural and non-structural adaptations of
questions and, to a very limited extent, national additions of items, categories and
questions. Each national study centre used national adaptation forms to document their
adaptations to their national survey instruments. The TALIS 2018 international study centre
(ISC) was responsible for approving all documented adaptations.
In preparation for the field trial and the main survey, decentralised translation took place.
This meant that each country was responsible for adapting and translating its own national
instruments from one of the two source versions (English or French). The translation
process required translation and translation review at the national level. For the main
survey, the ISC provided countries/economies with Word documents containing translated
questions from the field trial that could be used in unchanged form for the main survey.
During the translation verification process, overseen by the IEA Amsterdam, independent
language specialists at cApStAn Linguistic Quality Control (an agency specialising in
validating translations of international survey instruments) compared the translated
instruments side by side with the international versions. Upon completion of the
verification, the IEA Amsterdam returned the instruments, each of which contained verifier
feedback accompanied by a severity code (established by the IEA), to the TALIS 2018
NPMs and asked them to review the verifiers’ comments and improve the translations or
adaptations in line with the IEA’s guidelines for translation. From there, the NPMs
submitted the instruments to the ISC for verification of the layout, after which the NPMs
finalised their respective sets of instruments for data collection.
Most countries and economies taking part in TALIS 2018 administered the survey in
one predominant language, typically the language used throughout their entire education
system or at least understood by all respondents. Of the 48 participating
countries/economies, 9 administered the survey in more than 1 language (with the number
ranging from 2 to 5 languages). The ISC advised these countries/economies to involve
professionals familiar with more than one of these languages to review the translations and
to ensure equivalency across versions.
Participating countries/economies translated the principal and teacher questionnaires into
the languages listed in Table 4.1. The countries/economies that conducted online data
collection translated cover letters to be distributed to participants. The cover letters
contained information about the study, the web address for accessing the online
questionnaire and individualised user login information.
In general, each set of instruments underwent two rounds of translation verification – once
for the field trial and once for the main survey. However, two of the languages administered
during the main survey were not administered during the field trial. Norway initially
intended using only Bokmål but during the field trial decided to administer the instruments
in Nynorsk as well. This late decision, along with tight timelines, the fact that Bokmål was
the only language used in the regions participating in the field trial and the similarities
between Bokmål and Nynorsk, led to the national centre and the ISC agreeing that the
instruments in Nynorsk should undergo translation verification only during the main
survey. In Finland, the only language used during the field trial was Finnish, even though
the national centre had also prepared a Swedish translation. This translation was based on
the Swedish translation that Finland borrowed from Sweden. Due to the small number of
schools in Finland that would administer the instruments in Swedish, Finland elected not
to use the Swedish version during the field trial so as to prevent under-representation during
the main survey.
Several countries/economies used the English version of the instruments, notably England
(United Kingdom), Malta and the United States. Only two countries used the French
version – Belgium and France. These countries/economies all made national adaptations to
the instruments and submitted them for adaptation verification, language verification and
layout verification.
1. Note by Turkey: The information in this document with reference to “Cyprus” relates to the southern part of
the Island. There is no single authority representing both Turkish and Greek Cypriot people on the Island.
Turkey recognises the Turkish Republic of Northern Cyprus (TRNC). Until a lasting and equitable solution is
found within the context of the United Nations, Turkey shall preserve its position concerning the “Cyprus
issue”.
2. Note by all the European Union Member States of the OECD and the European Union: The Republic of
Cyprus is recognised by all members of the United Nations with the exception of Turkey. The information in
this document relates to the area under the effective control of the Government of the Republic of Cyprus.
3. For Finland, only Finnish was administered during the field trial.
4. For Norway, only Bokmål was administered during the field trial.
4.4. Adaptations
Adaptations were kept to a minimum but some were mandatory to ensure that the principals
and teachers in each country/economy received questions equivalent to those administered
to principals and teachers in all other countries/economies. Revisions included structural
and non-structural changes, as well as mandatory and elective adaptations.
Structural adaptations: The term “structural adaptation” referred to any adaptation
that altered the structure of the international source questionnaires by removing,
adding or splitting categories, by removing, adding or splitting national dimensions,
by removing or adding questions or by changing the width and range definition for
a question.
Non-structural adaptations: Non-structural adaptations referred to adaptations that
did not change the structure of the questionnaires, that is, neither added nor
subtracted questions, dimensions or categories to or from the instruments. The
adaptations typically involved adapting terms and phrases to fit the cultural context
of each country/economy. The adaptation of the term or phrase was more than a
pure translation because the international term or phrase was replaced by the
national term or phrase.
Mandatory adaptations: The international questionnaires contained terms or
phrases that needed to be adapted at the national level. These adaptations were non-
structural because they did not alter the structure of the instrument but adapted it to
fit a local context. To facilitate the mandatory adaptation process, the international
Table 4.2. Excluded TALIS 2018 principal questionnaire questions (before translation
verification)
Participating
country/ PQ-08 PQ-10 PQ-11 PQ-15 PQ-17 PQ-19 PQ-20 PQ-24 PQ-25 PQ-29 PQ-33 PQ-39 PQ-42
economy
Bulgaria H H B, C
Ciudad G–H D C, F
Autónoma de
Buenos Aires
(Argentina)
Croatia ISCED2: C C, F
A, D, E
ISCED3:
A–C
Cyprus1 CAT 5
Participating
country/ PQ-08 PQ-10 PQ-11 PQ-15 PQ-17 PQ-19 PQ-20 PQ-24 PQ-25 PQ-29 PQ-33 PQ-39 PQ-42
economy
Denmark ISCED3: F–H C
A
France D
Italy G, J X
Japan F
Korea E
Malta E
Spain C, M
Sweden D A–D C
United States D
Table 4.3. Excluded TALIS 2018 teacher questionnaire questions (before translation
verification)
Participating country/
TQ-14 TQ-15 TQ-20 TQ-24 TQ-29 TQ-35 TQ-36 TQ-50 TQ-55
economy
Belgium A–H
Flemish Community A–H
(Belgium)
Bulgaria X H X A–I
Ciudad Autónoma de A–H D
Buenos Aires (Argentina)
Croatia C
Israel F
Italy X
Japan ISCED1: A–H
G
ISCED2:
F
Netherlands ISCED1:
A–G, J–K
Shanghai (China) F
United States A–H G
Note: In the table, an “X” means that the respective country/economy excluded the question. For questions with multiple
dimensions, the listed letters refer to those dimensions of the respective question that were not administered in the
country/economy. For example, Croatia excluded dimension C of TQ-29. Where necessary, the information presented in the table
distinguishes between the different TALIS 2018 target populations (i.e. ISCED level 1, ISCED level 2, ISCED level 3 and TALIS-
PISA link). Unless stated otherwise, all information refers to the ISCED level 2 instruments.
All changes, selections and adaptations to the TALIS instruments were done with the goal
of creating an international database containing comparable data from all participating
countries/economies. NPMs noted all revisions to the instruments for the main survey in
an Excel document called a national adaptation form (NAF). NPMs received adequate
training in using the form and were required by the ISC to fill out a form for each language
and target population in which they intended to administer the instruments. Georgia, for
example, filled in a set of three NAFs because it administered TALIS in Azeri, Georgian
and Russian. The NAF ultimately contained the complete translation, adaptation and
verification history of each set of national instruments for every applicable target
population and it was, therefore, an integral part of the adaptation and translation processes
before and after international translation verification and, finally, layout review.
During preparation of the national instruments, the ISC asked the NPMs to submit the
adaptation forms at five key times (project “milestones”):
Step I, NAF approval: The ISC reviewed the proposed adaptations entered on the
NAFs. Terms and items used in TALIS 2013 had to be translated identically to
allow for trend analysis. The ISC released the TALIS 2013 main survey instruments
as a reference. The ISC also gave all participating TALIS 2018
countries/economies the International Standard Classification of Education 2011
mapping (UNESCO-UIS, 2012[1]) and asked them to use this for referencing the
ISCED levels.
Step II, translation verification: External language experts (from cApStAn)
reviewed the translated (updated) ISCED level 2 core instruments, referring to the
NAFs when relevant and commenting on any implemented adaptations. The
experts also verified translations of the cover letters and the NAFs for the other
international options (ISCED level 1, ISCED level 3 and the TALIS-PISA link).
Step III, layout verification: During this stage, the ISC compared the layout of the
national instruments with the international source versions in English or French and
noted any deviations from the international versions in the national instruments.
Step IV, online data collection (ODC) verification: The ISC prepared online
questionnaires for each participating country/economy in the respective
language(s). The online questionnaires included all adaptations to match the
national instrument structure. National study centres then used the online delivery
system to review the final online instruments.
Step V, codebook verification: The ISC created a national codebook for each
country/economy. Countries/economies were asked to check the structure of the
national codebook by entering one record for each questionnaire type.
TALIS 2018 offered countries/economies the opportunity to survey not only the
international target population, that is, the teachers and principals in schools providing
lower secondary education (ISCED level 2) but also the populations of teachers and
principals at ISCED level 1 and ISCED level 3 and those involved in the TALIS-PISA link.
Countries/economies that chose to survey teacher and principal respondents from more
than one of the international options needed to ensure that the different survey instruments
were consistent across the core and optional target populations. Deviations across target
populations were only permissible if the country/economy could justify them and if the ISC
approved those justifications.
As a further means of quality control, the ISC asked countries/economies to produce the
ISCED level 2 core version of the instruments in the predominant survey language and then
to use this version as the main blueprint for the international options and/or additional
language(s). This key process was ingrained in the survey operations to ensure that all
national materials were of high quality and consistent across the different instruments used
within one country/economy. The procedure applied to all steps of the national instrument
production that are outlined in this chapter.
The ISCED levels 1 and 3 instruments were not subject to a full translation verification.
Instead, revisions were recorded in the NAF, and verifiers then checked these changes. A
thorough consistency check during layout verification ensured only those adaptations to
the ISCED levels 1 and 3 instruments that the ISC had agreed to were implemented and
that the rest of the instruments matched the ISCED level 2 core instruments.
Preparation of the principal and teacher questionnaires for the TALIS-PISA link followed
the same procedure used for preparing the ISCED levels 1 and 3 questionnaires. However,
because the countries/economies administering the TALIS-PISA link had to replace the
ISCED-level information in those questionnaires with the notation “15-year-olds”, they had
to produce an extra set of questionnaires, even if they were also administering the survey
at an ISCED level where 15-year-olds were part of the population.
All participating countries/economies implemented the survey at the ISCED level 2 core,
15 implemented it at ISCED level 1, and 11 implemented it at ISCED level 3.
Nine participating countries/economies also took part in the TALIS-PISA link option (see
Table 4.4).
Table 4.4. List of the teacher and principal target populations in TALIS 2018 by country
ISCED level 2
Participating country/ economy ISCED level 1 ISCED level 3 TALIS-PISA link
(core)
Alberta (Canada) X X
Australia X X X
Austria X
Belgium X
Flemish Community (Belgium) X X
Brazil X X
Bulgaria X
Chile X
Ciudad Autónoma de Buenos X X X
Aires (Argentina)
Colombia X X
Croatia X X
Cyprus1 X
Czech Republic X X
Denmark X X X X
England (United Kingdom) X X
Estonia X
Finland X
France X X
Georgia X X
ISCED level 2
Participating country/ economy ISCED level 1 ISCED level 3 TALIS-PISA link
(core)
Hungary X
Iceland X
Israel X
Italy X
Japan X X
Kazakhstan X
Korea X X
Latvia X
Lithuania X
Malta X X
Mexico X
Netherlands X X
New Zealand X
Norway X
Portugal X X
Romania X
Russian Federation X
Saudi Arabia X
Shanghai (China) X
Singapore X
Slovak Republic X
Slovenia X X
South Africa X
Spain X X
Sweden X X X
Chinese Taipei X X X
Turkey X X X X
United Arab Emirates X X X
United States X
Viet Nam X X X X
The ISC advised NPMs to engage a minimum of two translators for each language in which
their respective countries/economies intended administering the survey. Translators needed
to have the scheduled language as their mother tongue, possess excellent knowledge of
English and be familiar with survey instruments.
The first of these translators, who was expected to be not only a language specialist but
also someone with a sound understanding of the country’s/economy’s cultural context,
translated the international English (or French) text of the instruments and manuals into the
national language. The second translator, known as the reviewer, was expected to possess
experience in the national educational context and to be familiar with the subject of the
study. This individual reviewed and commented on how appropriate the initial translation
was in terms of fitting the national educational context. This person also checked the
translation’s accuracy and readability. The country’s/economy’s NPM subsequently
reviewed the translation and the reviewer’s comments and made changes where he or she
deemed appropriate in the final document. This method meant that three independent
people compared the national versions against the original international source versions in
English or French.
The ISC reminded NPMs planning to divide up the translation work or to produce
translations in more than one language of the importance of ensuring consistency within
and across instruments. The ISC also encouraged the countries/economies that intended
producing the survey instruments in more than one language to engage professionals
familiar with all the languages as special reviewers so as to establish equivalence across
translations.
inaccuracies or word-for-word translations that were not appropriate within the national
language and context.
The translator’s role was to prepare a full translation of the core questionnaires and cover
letters so that these would provide the blueprints for the other international options.
Translators introduced the mandatory adaptations for the international options and, where
applicable, documented them in the NAFs. The translated texts needed to ensure
understanding and natural flow to the extent that anyone reading the texts could not tell if
they had originated in English or French. Guidance on language use during translation, as
outlined in the TALIS 2018 Survey Operations Procedures Unit 3, Instrument Preparation
(Main Survey), comprised the following:
The translated text should have the same register (language level and degree of
formality) as the source text.
The translated text should have correct grammar and usage (e.g. subject/verb
agreement, prepositions, verb tenses, etc.).
The translated text should not add text to or omit text from the source version unless
the ISC agreed to this.
The translated text should employ equivalent qualifiers and modifiers appropriate
for the target language.
The translated text should have the equivalent social, political or historical
terminology appropriate for the target language and used at this level of education.
Idiomatic expressions should be translated appropriately, not necessarily word for
word.
Spelling, punctuation and capitalisation in the target text should be appropriate for
the target language and the country/economy or cultural context.
When the individuals in the national study centres responsible for the translation work
completed their work, reviewers read the materials to ensure that the translations were
appropriate for teacher and principal respondents, were consistent with the field trial
version of the instruments and met the requirements of TALIS 2018. After the reviewers
had commented on the consistency and quality of the translations, the translators were
expected to integrate the changes into the instruments. If a translator and a reviewer
disagreed on the most appropriate translation, the NPM acted as an arbitrator and had the
right to make the final decision.
International translation verification was an important part of the TALIS 2018 technical
standards (specifically, Standards 4.18–4.27). The IEA Amsterdam co-ordinated the
translation verification process and engaged the services of native-speaking linguistic
verifiers through cApStAn Linguistic Quality Control, based in Brussels, Belgium. These
verifiers, experienced in balancing the cultural and national “appropriateness” of the target
version with “faithfulness” to the source version, provided expert feedback on
country/economy translations and adaptations. The ISC then asked the NPMs to carefully
review all verifier comments and suggestions and to implement those that improved the
questionnaire materials according to the IEA’s guidelines for translation, while also
ensuring that the translations retained the original meaning of the phrases. Although an
NPM always had the right to make final decisions regarding document content, he or she
had to describe and explain any major differences of opinions between national study centre
personnel and verifiers.
The ISCED level 2 core instruments and cover letters underwent full translation
verification. The ISC asked the NPMs to use the verified core instruments as the base from
which to develop the optional instruments and to make only approved adaptations to the
latter. Therefore, for the international options (ISCED level 1, ISCED level 3 and the
TALIS-PISA link), the additional entries in the NAFs were the only content to undergo
international translation verification. The IEA Amsterdam provided verifiers with the same
materials that the NPMs used when producing national translations.
During the main survey, translation verifiers received the international (English or French)
questionnaires in PDF format, which gave them an accurate preview of the intended format.
The translated questionnaires (for all ISCED levels and the TALIS-PISA link) and cover
letters (online data collection) were received in Word format, while the relevant NAFs were
received as Excel files. Verifiers used the “track changes” feature of Word to insert their
comments and changes directly into the national documents. Verifiers also documented and
recorded (in the form of comments) any deviations from the international source version in
the NAF and used the IEA severity code system to indicate the severity of the identified
error or issue.
Each verifier comment contained one of the following severity codes:
Code 1: Major change or error. Examples included the incorrect order of choices
in a multiple-choice item, omission of an item, incorrect translation resulting in the
answer being indicated by the item, an incorrect translation that changed the
meaning or difficulty of the text/item, and the incorrect order of items.
Code 1?: Not certain. The verifiers used a Code 1? when they were unsure of how
to correct a possible error or which intervention category to apply.
Code 2: Minor change or error. Examples included spelling errors that did not
affect comprehension.
Code 3: Suggestion for alternatives. The translation might be adequate, but the
verifier suggested different wording.
Code 4: Acceptable change. Examples included national conventions for
capitalisation and date format.
The ISC developed a survey activities questionnaire (SAQ) that it administered after the
field trial and again after collection of the main survey data. The questionnaire, designed
to elicit feedback from NPMs on their experiences administering TALIS 2018, included
questions on the translation and international translation verification processes. Responses
to the SAQ administered after the main data collection period indicated that the NPMs
found the international translation verification process benefited preparation of the main
survey.
All NPMs of the participating countries/economies answered the SAQ. Their answers
showed that the majority of the participants (29 of the 48) experienced no difficulties
translating the source questionnaires into national language(s) or adapting them to local
contexts. Most of the NMPs reported that after reviewing the feedback from the
As a final step during production of the national instruments and after completion of the
translation verification process, the ISC verified the layout of the paper versions and created
the online versions of the instruments. The ISC carefully checked the national versions of
the instruments against the international English or French versions and the documentation
in the corresponding NAFs. The aim of the layout verification was to ensure that the
national versions of the TALIS instruments looked, as much as possible, like the
international source versions and that the paper and online versions were, therefore, to the
greatest extent possible, equivalent.
The process used to verify the layout of the paper version was similar to the NAF approval
process. NPMs submitted all ISCED level 2 (core population), ISCED level 1, ISCED
level 3 and TALIS-PISA link instruments and, if applicable, the cover letters, to the ISC
for approval. The ISC then compared the ISCED levels 1 and 3 questionnaires, including
the teacher and principal questionnaires for the TALIS-PISA link, against the approved
ISCED level 2 questionnaires, which served as the new master versions.
ISC staff checked each questionnaire for font size, font changes, adjustment of cells,
response options, blank pages, word emphases, tracked changes, page breaks and
comments. If ISC staff found deviations from the documentation in the NAF, they adjusted
the paper versions and asked the relevant NPM to verify the correction of the mistake.
After paper layout approval, the ISC simultaneously implemented the instruments for all
ISCED levels and options in the IEA Online Survey System (OSS) and then checked each
online version of an instrument against its approved paper version. This practice helped
ensure that the instruments within one country/economy were the same, regardless of
whether they were administered on paper or on line. Visual checks were run using the same
standards and procedures as for verification of the paper layout. After finalising the online
files, the ISC asked each NPM to thoroughly check the files and report any mismatching to
the ISC.
Up to two rounds of checking and verification were needed for most of the language
versions of the instruments before they received final approval from the NPM.
The process involved in verifying the different international options was time-consuming
and many countries/economies were operating under a tight schedule. For these reasons,
the ISC invited those countries/economies that had elected to administer one or more of the
international options to work simultaneously on finalising the paper layout for these
options. The majority of NPMs (40 of the 48) said they found the instructions for layout
verification provided by the ISC helpful for preparing their national survey instruments. In
a few cases, NPMs detected some minor inconsistencies regarding spelling, punctuation or
layout mismatches after layout approval and prior to online system activation or printing.
The ISC changed these inconsistencies and informed the relevant NPM of the updated
version(s).
References
Notes
1
The TALIS School Co-ordinator Manual (internal document) contained detailed information about
the role and responsibilities of the school co-ordinator. The ISC provided two templates of the
manual in English to the NPMs, one template for online survey administration and the other for the
paper-based survey administration. The templates were to be translated and adapted by the NPM
and then distributed to the country’s/economy’s school co-ordinators.
2
TALIS 2018 set the standard that all participating countries/economies should implement 100% of
the agreed questions and items during the main survey data collection. However, systems were
offered the possibility of derogating specific questions and items if they considered they were not
suitable for their national context. Systems had to prepare a detailed explanation for each item they
wished to derogate and the request had to be approved by the OECD Secretariat and by the bureau
of the TALIS Governing Board. Only 2 of the 48 participating countries/economies requested a
derogation of items. Both of these requests were approved.
The international target population for the Teaching and Learning International Survey
(TALIS) 2018 consisted of schools providing ISCED level 2 education (deemed the core
survey of TALIS) as defined by the International Standard Classification of Education
2011, as well as their principals and their teachers. Participating countries and economies
could also opt to survey primary (ISCED level 1) and upper secondary (ISCED level 3)
teachers. This chapter covers the sample design prepared for the TALIS 2018 countries
and economies. It also reviews the sampling strategies and the nominal sample sizes. TALIS
2018 participants who also took part in the OECD Programme for International Student
Assessment (PISA) 2018 had the option of implementing TALIS in the schools that
participated in PISA 2018. This international option is called the TALIS-PISA link.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
5.1. Overview
This chapter covers the sample design prepared for the countries/economies participating
in TALIS. It also reviews the sampling strategy and the sample size. The chapter focuses
solely on the standard international (the “core” survey) sampling plan. Strategies for
estimation of population characteristics and of their sampling error are detailed in
Chapter 9, while Annex E provides characteristics of each national sampling plan.
The TALIS 2018 Sampling Manual (internal document) provides a more comprehensive
description of the survey design and its recommended implementation. The salient points
of the survey design appear in the section of this chapter on the participating countries’ and
economies’ core survey (ISCED level 2)1 samples.
Participating countries/economies could opt to survey ISCED level 1 and ISCED level 3
teachers. Australia, the Flemish Community of Belgium, Ciudad Autónoma de Buenos
Aires (henceforth “CABA”, Argentina), Denmark, England (United Kingdom), France,
Japan, Korea, the Netherlands, Spain, Sweden, Chinese Taipei, Turkey, the United Arab
Emirates and Viet Nam chose to survey ISCED level 1 teachers. Alberta (Canada), Brazil,
Croatia, Denmark, Portugal, Slovenia, Sweden, Chinese Taipei, Turkey, the United Arab
Emirates and Viet Nam chose to survey ISCED level 3. Participating countries/economies
were also offered an international option, which was to administer the TALIS teacher
questionnaire to a sample of PISA teachers in a sample of schools selected for PISA 2018.
This linking of TALIS data and PISA student achievement data at the school level is known
as the TALIS-PISA link. CABA (Argentina), Colombia, the Czech Republic, Denmark,
Georgia, Malta, Turkey and Viet Nam elected to participate in this school-level option.
The international sampling plan prepared for each of the TALIS 2018 teacher populations
was a stratified two-stage probability sampling design. This design meant that teachers
(second-stage units or secondary sampling units, abbreviated SSU) were randomly selected
from the list of in-scope teachers for each of the randomly selected schools (first-stage or
primary sampling units, abbreviated as PSUs).
The populations of interest comprised schools providing ISCED level 2 education, as well
as their principals and their teachers. TALIS adheres, for sampling purposes, to the OECD
international education statistics definition of a classroom teacher: “A classroom teacher is
defined as a person whose professional activity involves the planning, organising and
conducting of group activities whereby students’ knowledge, skills and attitudes develop
as stipulated by educational programmes” (OECD, 2004, p. p. 47[1]).
The TALIS programme of surveys (the ISCED level 2 core survey and the options for
ISCED levels 1 and 3 and the TALIS-PISA link) aims to cover all teachers of a given
ISCED level in a participating country/economy. Because the programme’s identification
of policy issues encompasses the classroom, the teacher, the school and school
management, all subjects being taught in a school are within TALIS’s scope. As such, the
programme’s sampling coverage extends to all teachers of an ISCED level and to the
principals of the schools in which these teachers are working.
Teachers at a given ISCED level are those who, as part of their regular duties in a target
school, provide instruction in programmes at that ISCED level. The TALIS populations of
interest also include teachers teaching a mixture of subjects at different levels in a sampled
school. This consideration applies no matter how much or how little teaching these teachers
are engaged in.
TALIS’s international target population restricts the survey to those teachers who teach in
“regular” schools and to the principals of those schools. Teachers teaching adults are not
part of the international target population and are therefore deemed “out of scope”; teachers
working with children with special needs are “in-scope” if they teach in regular schools.
However, when a school consists exclusively of these teachers, the school itself is said to
be out of scope. Teacher aides, pedagogical support staff (e.g. guidance counsellors,
librarians) and health and social support staff (e.g. doctors, nurses, psychiatrists,
psychologists, occupational therapists and social workers) are not considered teachers and
therefore cannot be part of TALIS.
Ideally, all the members of a target population should be admissible to sampling and data
collection, and this is the option that TALIS chose. As a consequence, the international
survey population (those who can be surveyed) is identical to the international target
population (those who should be surveyed).
For national reasons, participating countries/economies can choose to restrict the coverage
of their national implementation of TALIS to parts of the country/economy. For example,
a province or state experiencing civil unrest or an area that has recently been struck by a
natural disaster can be removed from the national target population to create a national
survey population. The TALIS sampling team asked participants to restrict these exclusions
to the greatest extent possible, for the reason given in the TALIS 2018 Sampling Manual:
“So as to maintain comparability and unbiasedness, exclusions should be kept to the strict
minimum and be justifiable. With only broad guidelines to help them, countries that
participated in the previous cycles of TALIS successfully managed to keep the proportion
of excluded teachers to less than 5%. A 5% threshold was thus adopted for this [third]
round of TALIS as an upper limit for the exclusion of teachers from the survey population”
(p. 11).
TALIS recognises that attempting to survey teachers in very small schools, that is, schools
with no more than three teachers at the ISCED level of interest, and those teaching in
schools located in geographically remote areas, can be costly, time-consuming and
statistically inefficient. Participating countries/economies can, therefore, excuse those
teachers from the TALIS data collection, thus creating a national survey population
different from the national target population. TALIS 2018 required the national project
manager (NPM) for each country/economy to document the reasons for exclusion, as well
as the size, location and clientele of each excluded school.
Ultimately, the TALIS 2018 samples of schools and teachers were selected from the
national survey population. Table 5.1 illustrates how the international and national survey
populations relate to one another. Note that Chapter 6 of this report covers the sampling of
teachers in the participating schools.
Annex D presents the national definitions of the TALIS 2018 target and survey populations
for each of the ISCED levels. NPMs provided this information on TALIS sampling forms,
templates of which can be found in Annex C.
During TALIS 2018, some teachers within a selected in-scope school were excluded from
the survey. They included:
teachers who were also acting as principals: no teacher data collected, but principal
data collected (labelled as NEXCL5 in Chapter 9)
substitute, emergency or occasional teachers: out of scope
teachers on long-term leave: out of scope
teachers teaching exclusively adults: out of scope
teachers in Cyprus,2,3 Iceland, Malta and the United Arab Emirates who took part
in the TALIS 2018 field trial so that they would not have to answer another TALIS
questionnaire (labelled as NEXCL6 in Chapter 9).
NPMs received detailed guidelines on how to apply these exclusion categories. Guidelines
could be found in the TALIS 2018 Survey Operations Procedures Unit 1: Sampling Schools
(internal document) or were provided during correspondence between Statistics Canada,
the international study centre and the interested participating countries/economies. In
addition, the TALIS 2018 School Co-ordinator Manual (internal document) provided
school co-ordinators with information on how to recognise different types of exclusion and
how to apply the correct codes.
NPMs were reminded that they were not to exclude teachers teaching at more than
one school. Instead, they were to record the number of schools in which these teachers were
working (see “weight adjustment for teacher multiplicity” in Chapter 9).
To allow for reliable estimation and modelling while also allowing for some degree of
non-response, TALIS 2018 set the minimum sample size at 20 teachers within each
participating school and required countries/economies to draw a minimum sample of
200 schools from the national population of in-scope schools. The nominal sample was,
therefore, a minimum of 4 000 teachers.
Teachers from the same school tend to share opinions and behave in similar ways more so
than teachers from different schools, cities or provinces in a given country/economy. This
tendency for two teachers from the same school to be “more alike” than two teachers from
different schools is called the “clustering effect” and it is measured, in single-stage designs,
by the intra-cluster correlation. In essence, the stronger the intra-cluster correlation, the
lower the number of sampled teachers needed from a school because one responding
teacher becomes a good predictor of the other teachers in that same school. In other words,
in a sample of 20 teachers from the same school, there are, in a sense, fewer than 20 original
data points. This outcome is a manifestation of the clustering effect or design effect, and
the larger the cluster, the larger the loss.
Those engaged in the preparation work for TALIS 2013 used an intra-cluster correlation
value of 0.3 as a working hypothesis, on the supposition that teachers are as homogeneous
as their students, this supposition accorded with the design of TALIS 2008. The team that
worked on the design of TALIS 2018 adopted the TALIS 2013 design, in conformity with
the Terms of Reference. The loss in sample size due to clustering, when added to the losses
due to non-response, reduced the nominal sample of 4 000 teachers to an effective sample
of approximately 400 as depicted in Table 5.2. Thus, the nominal sample of 4 000 teachers
obtained by the complex sampling design was equivalent to a simple random sample of
433 teachers.
Schools a 200
Teachers per school b 20
Total number of teachers c=a×b 4 000
School response rate d 75%
Teacher response within school e 75%
Overall response rate f=d×e 56%
Net number of responding teachers g=c×f 2 250
Intra-cluster correlation h 0.30
Design effect (deff) deff = 1 +{(e ×b)-1}×h 5.2
Effective sample = g / deff 433
The precision that is expected from the sample of 20 teachers in 200 schools is equivalent
to that of a simple random sample of 433 teachers selected from the (often unavailable)
national list of teachers. The expected margin of error for a simple random sample of this
size is ± (1.96) (1/√433) = ± 9.4%. Evidence from TALIS 2008 and TALIS 2013 shows
that, in most countries, clustering was not as great as anticipated. Hence, the achieved
precision in most countries and for most statistics was better than the expected 9.4%.
However, the requirements for the nominal sample for 2018 remained at the original level
to allow for easier tabulations at subnational levels and for more robust secondary analyses.
Participating countries/economies could choose to augment their national sample by
selecting more schools. Alternatively, they could select more teachers to increase the
within-school sample and thereby counterbalance the effect of selecting too many schools
with too few teachers.
The sampling team reduced the sample size requirement for some participating
countries/economies because of the smaller number of schools available for sampling (see
Annex E, which presents the characteristics of the national samples). In the few cases where
the average number of teachers in the schools was lower than the number given in the
international plan, the sampling team asked for the number of schools sampled to be
increased to maintain a minimum total number of participating teachers.
5.5.2. Stratification
The international sampling plan did not require stratification of the schools or of the
teachers within the selected schools. The sampling team invited participating
countries/economies that chose to implement some form of stratification (in order to answer
national requirements) to discuss their strategy with them.
Stratification could be done explicitly (whereby a fixed portion of the total sample is
allocated to the stratum) or implicitly (the variable is used to sort the sampling frame before
sample selection, thus giving, on average, a proportional representation of the implicit
strata in the sample).
In instances where explicit stratification was used, the participating country/economy and
the sampling team together determined the sample allocation scheme.
In most cases, stratification resulted in a combination of some or all of the details relating
to geography, source of financing, type of educational programme and school size.
Annex E (Tables 5.7 to 5.9) provides these details for each participating country/economy
and each ISCED level in which they participated.
by implicit strata and measure of size (MOS) prior to sampling. Sampling frames were
always sorted by MOS prior to sampling, whether or not stratification was applied. Sorting
by MOS was done in a serpentine manner, which meant alternating increasing order and
decreasing order so that adjacent schools would be of similar sizes even across strata. This
approach is useful when creating replication zones for estimation of sampling error (see the
section in Chapter 9 on creating replicates for balanced repeated replication).
The mechanics of systematic random sampling with PPS can be described as follows. Let
M be the total MOS in an explicit stratum, let mi be the MOS for school i in the explicit
stratum and Mi be the cumulative sum of the school sizes up to and including school i, and
let n be the number of schools to be sampled from that explicit stratum. A sampling step k
is then computed as quotient Mn, and a starting point d is drawn at such that 1 d < k+1.
The sample is selected by walking steps of fixed length k along the (ordered) sampling
frame. As evident in Table 5.3 below, the point at which the step lands points to the school
to be added to the sample.
Whenever possible, the sample selection programme selected two replacement schools for
each sampled school: the school just above and the school just below the selected school
on the sampling frame sorted by MOS. The replacement schools had to come from the same
explicit stratum as the sampled school. The sampling team advised the use of this strategy
to help maintain the sample size and minimise the non-response biases that can occur when
schools with characteristics similar to those of the non-responding schools are used.
Schools selected for the original sample could not be selected as a replacement school.
To simplify and speed up the sampling process, the sampling team selected all samples of
schools.
At the end of school selection, the sampling team sent each participating country/economy
a copy of its school sampling frame, in which the selected schools were identified (marked
“S” for the original sample and marked “R1” and “R2” for the replacement schools) and
then given a standardised TALIS school identification number.
Table 5.3 illustrates how an ordinary spreadsheet can be used to implement systematic
random sampling with PPS. In this illustration, explicit stratum “A” consists of 12 schools
with a total MOS of 209 teachers. The sample needed from this stratum is n = 3 schools;
the sampling step k (209 ) = 69.7. Suppose that the random start is d = 49. The jth school
selected is then such that Mj-1 < d + (j-1) × k Mj, with M0 = 0 and j = 1, 2, 3. Here, for the
first selection, j = 1 and the pointer is 49 + (1-1) × 69.7 = 49. If j = 2, the pointer is at 49 +
(2-1) × 69.7 = 118.7 (rounded to 118), and finally the pointer is at 118.7 + 69.7 = 188.4
(rounded to 188). If available, replacement schools (the schools immediately before and
after a selected school) are assigned automatically; note that School 12 has no second
replacement.
The International Association for the Evaluation of Educational Achievement (IEA)
Hamburg provided each participating country/economy with the IEA Windows Within-
School Sampling Software (WinW3S) to help them create the sampling frames and sample
selection of teachers and to ensure compliance with the sample design and with furnishing
complete documentation.
Annex E presents the size of the sample of schools and of teachers for each participating
country/economy.
Selections
National school Implicit Cumulative Sampling
Explicit stratum MOS mi Pointer and
ID stratum MOS Mi step
replacements
1 A 1 10 10
2 A 1 12 22
3 A 1 15 37 R1
4 A 1 17 54 1 49 S
5 A 2 20 74 R2
6 A 2 18 92
7 A 2 16 108 R1
8 A 2 16 124 2 118 S
9 A 3 15 139 R2
10 A 3 17 156
11 A 3 26 182 R1
12 A 3 27 M = 209 3 188 S
The following three tables give an overview of the sampling plan for each participating
country/economy.
Table 5.4 covers the countries/economies that participated in ISCED level 2, Table 5.5
those countries/economies that participated in ISCED level 1 and Table 5.6 those
countries/economies that participated in ISCED level 3.
Reference
OECD (2015), “ISCED 2011 Level 2: Lower secondary education”, in ISCED 2011 [2]
Operational Manual: Guidelines for Classifying National Education Programmes and
Related Qualifications, OECD Publishing, Paris, https://dx.doi.org/10.1787/9789264228368-
6-en.
OECD (2004), OECD Handbook for Internationally Comparative Education Statistics: [1]
Concepts, Standards, Definitions, and Classifications, OECD Publishing, Paris,
https://doi.org/10.1787/9789264279889-en.
Notes
1
As defined by the International Standard Classification of Education 2011 (OECD, 2015[2]),
2
Note by Turkey: The information in this document with reference to “Cyprus” relates to the
southern part of the Island. There is no single authority representing both Turkish and Greek Cypriot
people on the Island. Turkey recognises the Turkish Republic of Northern Cyprus (TRNC). Until a
lasting and equitable solution is found within the context of the United Nations, Turkey shall
preserve its position concerning the “Cyprus issue”.
3
Note by all the European Union Member States of the OECD and the European Union: The
Republic of Cyprus is recognised by all members of the United Nations with the exception of
Turkey. The information in this document relates to the area under the effective control of the
Government of the Republic of Cyprus.
4
Requirements for ISCED level 1 and ISCED level 3 were identical to those imposed for ISCED
level 2; the nominal sample size for the TALIS-PISA link was set at 150 schools, which is the PISA
requirement.
5
Only one replacement school was selected for the field trial to minimise the overlap with the sample
for the main survey. Schools for the field trial of the TALIS-PISA link component were selected by
convenience.
This chapter summarises the procedures and outcomes of the TALIS 2018 field operations.
It focuses on the areas of work within national centres, which included contacting schools,
performing the within-school sampling, and monitoring the collection of data for the
International Standard Classification of Education (ISCED) level 2 core and eventual
additional international options of the TALIS 2018 main survey.
The chapter also references the materials and software that the international study centre
(ISC) gave all national study centres.1
The overall administration of TALIS 2018 in all participating countries/economies was
very successful. No mayor obstacles were encountered and the field-operation procedures,
as defined and communicated by the ISC, were met.2
Although the procedures described in this chapter focus on the administration of the ISCED
level 2 core survey, they also apply to the international ISCED level 1, ISCED level 3 and
the TALIS-PISA link options. The ISC asked all participating countries/economies to follow
the standards and procedures and made it clear that deviations from these would only be
allowed in certain cases and if discussed and agreed with the TALIS Consortium and the
OECD.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
During all phases of the survey, the national study centres adhered to the standardised
procedures prepared by the international study centre (ISC) and its consortium partners.
These procedures were outlined in the following documents, which the ISC released to the
national study centres before the field trial and then updated for the main survey (except
for the technical standards). To note, apart from the technical standards, these are all
internal documents:
TALIS 2018 Technical Standards (see Annex B): The ISC released the final version
prior to the main survey (MS) after the TALIS Governing Board (TGB) approved
the standards during its third meeting in June 2017. The technical standards covered
ten main topics,3 ranging from survey ethics, confidentiality and survey operations
to quality observation and data management.
TALIS 2018 National Project Managers’ Manual: The purpose of this manual was
to provide national project managers (NPMs) with an overview of TALIS, detail
the tasks NPMs and national study centres were to carry out, and give information
about key milestones and deliveries. The ISC provided detailed information about
national instrument production and survey operations in separate guideline
documents.
Survey Operations Procedures Unit 1: Sampling Schools: Statistics Canada
prepared this manual, which defined the target population of teachers for all of the
ISCED levels and the TALIS-PISA link. The manual also described how to prepare
and implement a national sampling plan, how to prepare the school sampling frame
and how to select the school sample. The manual’s annexes provided thorough
instruction on how to handle the samples for the different international options.
Survey Operations Procedures Unit 2: Working with Schools: As indicated by its
title, this manual addressed guidelines for obtaining and increasing co-operation of
schools, with the emphasis placed on public relations strategies. It also described
how to adapt the School Co-ordinator Manual to the national context and explained
details of instrument shipping and quality-control measures.
School Co-ordinator Manual: This manual was intended for school co-ordinators
(SCs). Each SC was the main contact person for the national study centre in each
school. The person fulfilling the role of SC was often a teacher or a principal of the
school participating in TALIS. The ISC released one template of the School Co-
ordinator Manual for participating countries/economies doing paper-only
administration and one template for those administering the survey solely on line
or in a mixed mode (paper and online administration). The manual described, in
detail, the steps for listing and tracking teachers and for organising the survey
administration on site. NPMs were responsible for translating the manual into their
survey administration language(s) and for adding national information where
necessary. Responsibility for translations and adaptations rested solely with the
NPMs. The ISC asked international quality observers (IQOs) to make sure the
NPMs used the correct template (see Chapter 7 for more details on quality
assurance).
Survey Operations Procedures Unit 3: Instrument Preparation: This manual
provided the national study centres with instructions on how to produce
internationally comparable national versions of the released international survey
A third software package, the IEA Online Survey System (OSS), helped the ISC prepare the
questionnaires for online administration. During this process, the ISC asked the
participating countries/economies to review the prepared online questionnaires via the
Internet. The web-based monitor component of the OSS allowed national centres to audit
participation in real-time and to follow up those schools that returned incomplete
questionnaires or did not return questionnaires.
During meetings with the NPMs, the ISC described and explained the field operation
procedures outlined in the manuals and guidelines and provided guidance on how to use
the software packages. Representatives of the national study centres also had to participate
in hands-on training sessions to practise the correct handling of the software and some
procedures. If any queries or concerns regarding procedures, guidelines or software
emerged during implementation, the ISC was available to support the national centres and
help them find appropriate solutions.
Due to the tight timelines for the Southern Hemisphere countries/economies, the ISC
delivered the School Co-ordinator Manual earlier than scheduled to those national study
centres that requested it. The ISC also released instruments and some manuals and forms
earlier than originally scheduled, that is, on 1 August 2017 rather than on 15 August. In
addition, the ISC treated Australia, Brazil, Chile, Ciudad Autónoma de Buenos Aires
(Argentina), Colombia, Korea, New Zealand and Singapore as high-priority and scheduled
exclusive time during August and September 2017 for verifying their survey instruments.
Data collection for Northern Hemisphere countries was mainly from March to May 2018
(with some participants starting early in January and February and some extending into
July 2018). In the Northern Hemisphere, schedules and procedures for Denmark, Turkey
and Viet Nam were particularly complex because these countries decided to administer all
of the international options. That decision resulted in 16 instruments across the options: 4
teacher questionnaires, 4 principal questionnaires, 4 teacher cover letters and 4 principal
cover letters.
Spain administered the survey in 5 different languages, which resulted in 20 instruments
(teacher questionnaire, principal questionnaire, teacher cover letter and principal cover
letter for each language). Spain also decided, in early 2017, to join the ISCED level 1
option, which meant it had to conduct a late field trial (FT) in September/October 2017.
Because administration of the Spanish FT coincided with preparation for the Spanish MS,
this work was particularly challenging for all parties involved. France, meanwhile, had to
administer a second FT in October/November 2017 in order to trial a set of politically
sensitive questions that it had not been able to administer during the original FT in the
spring of 2017.
Saudi Arabia and South Africa had very demanding schedules as well. South Africa
decided to join TALIS in September 2017 and therefore had to conduct its FT in
March 2018 and administer its MS in August-October 2018 in line with a Southern
Hemisphere schedule. Saudi Arabia administered its FT in March 2018, followed shortly
after by administration of its MS in May 2018.
Then there were participating countries/economies for which instrument preparation was
relatively straightforward and timelines were comfortable. These countries/economies,
(e.g. Estonia, Hungary, the Russian Federation) opted to administer only the ISCED level 2
core survey and in one national language only.
Statistics Canada sent each NPM a selected school sample based on the sampling frame the
NPM had already submitted (see Chapter 5 for more details on school sampling). In order
to achieve the highest possible participation rates at the school level, Statistics Canada
sampled two replacement schools (assuming such schools were available) in addition to
each originally sampled school.
Once NPMs received their sample, the national study centres began contacting the
designated schools. National study centres usually adhered to the technical standards and
only contacted their first replacement school if one of the originally sampled schools
declined participation. If this replacement school also refused participation, NPMs
approached their second replacement school.
Two participating countries (the Netherlands and the United States) applied another process
so as to achieve the required participation rates, namely contacting all schools at the same
time. If the original school and the replacement school both agreed to participate in the
survey, the replacement school was marked as a national school. However, because this
procedure represented a deviation from Technical Standard 5.5, national schools were not
included during computation of the participation rates, conducted as part of the data
adjudication process. Further mention of the adjudication procedure appears in the
discussion in Chapter 10 on data adjudication.
National study centres that administered the survey in 1 or more of the international options
generally had to handle, in parallel, up to 750 schools (including the 200 for the core
survey).5 These centres, therefore, had to plan their resources carefully to meet the
requirements of the multiple tasks associated with this complex survey design.
Most of the participating countries/economies asked each school to nominate a school
co-ordinator to be responsible for carrying out all TALIS-related tasks within the school.
In many cases, the school co-ordinators were principals or another school management
team member. Nearly 50% of the national study centres said that a member of the
management team other than the principal took on this responsibility. Nearly 40% of the
study centres noted that teachers filled the role of school co-ordinators. Only one national
study centre hired an external agency and only one centre hired external staff (e.g. retired
principals or experienced assessment co-ordinators) to complete the tasks.
Close co-operation between school co-ordinators and national study centres was crucial
during all steps of teacher listing, teacher sampling and survey administration. To facilitate
smooth communication, several countries/economies established hotlines, special email
accounts, or websites and online platforms.
6.3.1. Identification number, teacher listing forms and teacher tracking forms
Information about teachers was gathered through the teacher listing form (TLF) and teacher
tracking form (TTF; see Annex F for more information on these two forms). National study
centres used the IEA WinW3S software to produce the forms. WinW3S created
hierarchical four-digit identification numbers that uniquely identified the sampled schools
within each participating country/economy. This number was also the identification code
assigned to the person answering the principal questionnaire. Teacher identification codes
were derived from the school codes by adding two additional digits at the end of the school
identifier, a process that created a hierarchical link between schools and teachers.
In accordance with the instructions in the School Co-ordinator Manual, school
co-ordinators listed each eligible teacher and his or her name, followed by a sequential
number, exemption information, year of birth, gender and main subject domain. Nearly
one-third of the NPMs reported that data protection/confidentiality laws and rules
prevented them from providing teachers’ names. These participating countries/economies,
therefore, used only numbers or codes.
Schools could choose up to a maximum of five main subject domains: language, human
sciences, mathematics and science, other, and not specified. Classifying teachers into the
domains was sometimes a demanding task, requiring close co-operation between the school
co-ordinators and their respective national study centres. Although the TALIS core survey
targeted ISCED level 2 teachers, not every teacher teaching at this level was within scope.
Out-of-scope teachers included teachers entirely devoted to adult education; substitute,
emergency or occasional teachers; teachers on long-term leave; teacher aides; pedagogical
support staff; and health and social support staff. Teachers who were also the school
principal, as well as teachers who took part in the FT, were exempted from participation
but they still had to be included on the TLF (for more details on school sampling, see
Chapter 5).
The national study centres entered information from the TFLs into WinW3S and then drew
the random within-school teacher sample of 20 teachers per school.6 After completion of
the within-school sampling, WinW3S created TTFs that listed all sampled teachers. The
national study centres sent the TTFs to schools so that school co-ordinators knew which
teachers should receive the instruments.
The TTFs were also used to monitor the participation status of the sampled teachers and
therefore included teacher names, teacher ID, year of birth, gender, main subject domain
and teacher questionnaire mode (online or paper). Each TTF furthermore contained a
column that allowed the national study centres to document the teacher questionnaire return
status for paper administration and a column in which the centres could enter data
availability information from the online data collection. Because the form provided school
co-ordinators with a roster of selected teachers, they could also use the form to identify
which teachers within schools they needed to follow up via email or telephone.
The national study centres sent copies of the TTFs, each of which included teacher IDs but
not teacher names, to the ISC together with the survey data. Because the names on the TTFs
could be cut off, all names were kept confidential. Annex F contains a blank TFL and TTF.
Figure 6.1. Responsibilities of national study centres and the school co-ordinator during
survey administration
SURVEY ADMINISTRATION
The ISC asked each participating country/economy to select its own timeframe for survey
administration, but within the internationally prescribed time period at the end of the school
year. For Southern Hemisphere countries/economies, the internationally prescribed time
for the MS was between 1 October and 15 December 2017. However, some
countries/economies started in September, while others extended their data collection into
January 2018. For the Northern Hemisphere countries/economies, the prescribed time was
between 1 March 2018 and 31 May 2018, but some countries/economies started in January
and February, while others extended into July 2018 as an exception. National study centres
had to discuss any deviations from these periods with the ISC and gain the ISC’s approval
for them. The only deviations exempt from this process were those for South Africa, which
operated under a shifted survey schedule.
Table 6.2 documents the data collection windows as they were originally planned, as well
as changes to them. The table also highlights the dates that fell outside the internationally
prescribed time periods. Some national study centres had to exceed the data collection
window in order to achieve the necessary participation rates. They did this by starting the
data collection earlier or by finishing it later than originally specified. The data collection
windows, therefore, ranged from 13 days to 5 months.
While 15 countries/economies asked for start or end dates that were not in the prescribed
periods, most of these dates were still within the same school year. The only participating
countries/economies in which data collection shifted to a new school year were Australia
and Denmark (see Chapter 10 for details).
Table 6.2. Originally planned and actual data collection windows for ISCED levels 1, 2 and 31
1. The TALIS-PISA link was administered together with all the other ISCED levels in TALIS and in parallel
or shortly after the PISA main survey in the Northern Hemisphere countries/economies. The Southern
Hemisphere countries/economies (Australia and Ciudad Autónoma de Buenos Aires [Argentina]) administered
PISA and therefore the TALIS-PISA link on a shifted schedule between September and November 2018.
2. Note by Turkey: The information in this document with reference to “Cyprus” relates to the southern part of
the Island. There is no single authority representing both Turkish and Greek Cypriot people on the Island.
Turkey recognises the Turkish Republic of Northern Cyprus (TRNC). Until a lasting and equitable solution is
found within the context of the United Nations, Turkey shall preserve its position concerning the “Cyprus
issue”.
3. Note by all the European Union Member States of the OECD and the European Union: The Republic of
Cyprus is recognised by all members of the United Nations with the exception of Turkey. The information in
this document relates to the area under the effective control of the Government of the Republic of Cyprus.
Bolded dates highlight dates that fell outside the internationally prescribed periods.
The end of the school year was purposely selected for administering the survey to guarantee
comparability of collected data. During this period, principals and teachers were free to fill
in the questionnaires whenever they chose. The overall target was 100% within-school
participation. A school was considered to be a participating school if at least 50% of the
sampled teachers returned their completed questionnaires.
To maintain high survey standards, the ISC expected the national study centres to run a
national quality control programme. The ISC provided a manual template for national
quality observers (NQOs); however, NPMs could elect to arrange their own programme.
NQOs could perform their national quality observations partly on the phone and partly
during school visits, but the ISC also required each NQO to personally visit at least
ten schools.
Some participating countries/economies chose national study centre staff to carry out the
NQO programme, whereas others appointed external personnel, such as researchers
interested in large-scale assessments, representatives of government agencies involved in
education and retired principals. After the TALIS MS had been administered, NPMs
reported the outcomes of the national quality control work in the survey activities
questionnaire. Chapter 7 provides a discussion of these outcomes. The IEA Amsterdam
was responsible for organising the international quality observation work; their role in this
regard is also outlined in Chapter 7.
Monitoring the administration of the survey was a demanding task for the school
co-ordinators, especially if data protection laws prevented them from using the online
monitor of the OSS or if teachers could send completed paper questionnaires directly to
their national study centre. In these instances, school co-ordinators relied on national study
centres to inform them of the need to follow up pending questionnaires.
National study centres monitored completion of the online questionnaires (see Chapter 8
for more details),and NPMs communicated completion status to school co-ordinators.
Countries/economies were free to manage this procedure according to their needs. Most
national study centres preferred regular email and/or telephone exchanges between the
school co-ordinators and themselves. Other centres either contacted respondents directly
using email or instant messaging, set up banners on websites, prepared reminder leaflets
for teachers and schools, asked union members to call schools or created national TALIS
websites where school co-ordinators could log on individually to access all necessary
information. Some national study centres granted school co-ordinators access to the
monitor so that they could organise the follow-up procedures themselves.
The monitoring work also included a participation-rate estimation tool that kept data
managers up to date on their current participation rate according to the already returned
paper questionnaires or submitted online questionnaires.
After the survey had been administered, each national study centre exported the
questionnaire data availability status from the OSS. National study centre personnel then
imported this participation information, as well as the participation information from the
DME with respect to administration of the paper questionnaire, into WinW3S, a practice
that enabled the national study centres to verify the participation status of each sampled
respondent. The ISC told each national centre that it was mandatory for them to verify
participation before they submitted data to the consortium, a process that all TALIS
countries/economies completed successfully.
The major tasks for NPMs immediately after administration of the TALIS MS included
retrieving and collating the materials from schools and verifying their integrity. On
receiving survey materials from the schools, NPMs:
checked that they had received complete and appropriate cover
letters/questionnaires for every teacher listed on the TTF
verified that all identification numbers on all cover letters/questionnaires were
accurate and legible
checked that the participation status recorded on the TTFs matched the availability
of questionnaires, the information on the paper questionnaires and the information
in the online monitor
followed up those schools that did not return all the survey materials or for which
forms were missing, incomplete or otherwise inconsistent.
National study centres recorded all necessary information about schools, principals and
teachers, as well as the return status of the questionnaires, in WinW3S. NPMs then
organised the paper questionnaires and corresponding forms for data entry (see Chapter 8).
Notes
1
The international study centre (ISC) had prepared software packages for sampling, online data
collection, data entry and data processing to fit the needs of TALIS. The ISC requested that all
participating countries/economies use solely the IEA software to perform the survey; no exceptions
were allowed.
2
With the survey activities questionnaire (SAQ), administered after the field trial and again after the
main survey data had been collected, the ISC collected NPMs’ feedback on their experiences
administering TALIS 2018. The outcomes of the field trial activities questionnaire were considered
in terms of improvements to the main survey’s procedures and the software in use. The outcome of
the main survey questionnaire was incorporated into this chapter.
3
The ten topics included: survey ethics and planning; communication; sampling design, weighting
and adjudication; instrument adaptation, translation and verification; school co-operation and
within-school sampling; data collection and participation monitoring; observing the quality of data
collection; data capture of paper instruments, verification, submission and management;
confidentiality, security and preparation of the international database; data analysis and reporting.
4
If a participating country/economy administering one or more of the options was struggling with
low participation rates during the survey, the national study centre knew it had to first focus on the
core survey.
5
Exceptions occurred for small countries, where the sample size was reduced due to local
circumstances.
6
National study centres were given the opportunity to sample more than the recommended
20 eligible teachers per school, if desired.
This chapter describes the quality control programme developed and implemented for
administration of the Teaching and Learning International Survey (TALIS) 2018. Assuring
the quality of the data collection was a three-part process, comprising: (1) an international
quality control programme overseen by the IEA Amsterdam and designed to document the
procedures for survey preparation and administration during the main data collection; (2)
a national quality control programme carried out by national study centres; and (3) a
follow-up online survey activities questionnaire that asked national project managers to
comment on the implementation of the TALIS main survey procedures in their respective
countries.
Quality control in the survey administration was extremely important for ensuring valid
comparisons of teacher and principal survey results across countries/economies. It helped
document not only the quality of survey administration and adherence to technical
standards and to standardised survey administration guidelines and procedures in each
participating country/economy but also issues that could influence the quality and
comparability of the data.
Because the information collected by the international quality observers serves as evidence
for further analysis and improvements and reveals any critical and significant process-
related issues, the school visit record and the survey activities questionnaire provided data
on key components of the survey process and of national project managers’ experiences in
conducting TALIS 2018. The international quality observers gained the impression from
their observations that most countries/economies closely followed survey procedures.
Available evidence suggests that teacher respondents generally had few problems with the
survey. Any issues or concerns regarding survey administration procedures and problem-
solving were shared between the national study centres and the TALIS International
Consortium. The quality of the implementation of the TALIS 2018 data collection
procedures was well documented, mostly by quality observers at the international level.
Data reflected stable and consistent data collection processes across participants. In
addition, the importance of effective communication in facilitating international project
work and supporting project completion was, once again, highlighted, setting the stage for
attainable success during future TALIS cycles.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
During TALIS 2018, considerable effort went into developing standardised materials and
procedures that ensured the data collected in each country/economy were internationally
comparable to the greatest extent possible. More specifically, this standardisation process
ensured that survey materials were administered to participants under comparable survey
conditions for comparable analysis, across all countries/economies and languages and for
each mode of administration (i.e. online and/or paper data collection). Quality control was
implemented at different levels and stages during instrument production, administration,
and data entry and processing in order to document the extent to which each
country/economy implemented the standard operating procedures.
This chapter describes the outcomes of the quality control activities conducted during the
main data collection. Quality control consisted of three major parts:
The IEA Amsterdam designed and managed a standardised, international quality
control programme of school and national centre visits carried out by international
quality observers (IQOs). International quality control was implemented for the
main survey only.
The data from the quality control activities for TALIS were augmented by the
responses of national project managers (NPMs) to a survey activities questionnaire
(SAQ) administered online and conducted after administration of the main survey.
The questionnaire elicited information about the NPMs’ experiences in preparing
for and conducting the TALIS 2018 data collection and the extent to which
everyone involved followed procedures and guidelines. It also provided NPMs with
an opportunity to provide feedback about all aspects of survey administration.
Information pertaining to the national quality control programmes was also
reported in the SAQ and is presented later in this chapter.
The TALIS International Consortium required the NPMs to implement national
data collection quality observations (i.e. a national quality control programme made
up of school visits) during both the field trial and the main survey. The design of
the national quality observation programme was nevertheless at the discretion of
each participant. As a member of the consortium, the IEA Amsterdam provided a
national quality observer manual template that countries/economies could adapt to
suit their national contexts and use as a basis for training their national quality
observers (NQOs).
During TALIS 2018, countries/economies could administer three international options in
addition to the core ISCED level 2 population, namely ISCED level 1, ISCED level 3 and
the TALIS-PISA link. The sampled populations (and instruments) applicable to
participants featured in the design of the quality control programme and were
proportionally included in the schools selected for international quality observations in
each country/economy. The TALIS 2018 Consortium advised NPMs to adopt a similar
approach with respect to national quality control.
As part of the OECD TALIS 2018 quality assurance process, the IEA Amsterdam
established a standardised international quality control programme to document data
collection activities in the participating countries/economies.
IQOs were also required to collect the following documents from the national centre in
their country/economy:
national versions of the TALIS 2018 School Co-ordinator Manual (one for each
surveyed population, administration mode and language)
national survey instruments
a teacher listing form and a teacher tracking form for each school selected for
international quality observation.
IQOs had three main responsibilities with respect to the international quality control
programme. Their first task involved visiting the national study centre to interview the
country’s/economy’s NPM, collect national TALIS materials and select 20 schools to be
visited. Since a structured interview with the NPM would ensure a better understanding of
how the international procedures had been adapted to suit national contexts, the IEA
Amsterdam provided IQOs with an NPM interview outline and question template.
Second, IQOs were expected to complete a translation verification report (one report per
language of administration; a maximum of two languages). Each participating
country/economy had to translate and/or adapt the TALIS materials to the national situation
and submit all versions of their instruments1 (in all common languages) to the IEA
Amsterdam for independent international translation verification. The IEA Amsterdam
asked IQOs to review the national instruments and comment on the use of specialist advice
from the verifier regarding national translations of the international source instruments into
the official language(s) of instruction. They also had to assess and document consistency
across the national instruments used to survey the core target population and any optional
populations, and to compare the national version of the School Co-ordinator Manual with
the international templates in order to verify alignment (proper adaptation) and determine
if the TALIS 2018 guidelines produced by the international study centre (ISC) had been
followed.
The third duty, irrespective of the options (ISCED level 1, ISCED level 2, ISCED level 3
and the TALIS-PISA link), required IQOs to visit a total of 20 sampled schools to interview
the school co-ordinators (SCs) about TALIS activities leading up to and including the
distribution of materials, and to record their observations and interview responses in the
school visit record for each of the schools. The interview times ranged from approximately
30 minutes to no more than 60 minutes. IQOs were instructed to allow extra time for
questions that required a more careful explanation.
During their fieldwork, IQOs were required to report their progress to the IEA Amsterdam
and to advise of any issues that had arisen. As a result of their duties, IQOs had to send two
sets of deliverables to the IEA Amsterdam at different stages of the programme. The
deliverables included the following materials and documents:
a copy of the completed NPM interview outline and question template
a copy of the translation verification report per language of administration
a checklist documenting the materials collected from the NPM
a list of all visited schools
copies of all 20 completed school visit records
teacher listing forms and teacher tracking forms for every visited school
In co-operation with the NPM, the IQO in each country/economy selected 20 of the
sampled TALIS 2018 schools. For countries and economies that chose one or more of the
optional survey population(s), the number of school visits was distributed according to the
plan indicated in Table 7.1. If, for example, a country/economy administered two extra
international options in addition to core ISCED level 2, its IQO had to visit seven schools
for that ISCED level, seven schools for one of the international options (at the IQO’s
discretion) and six schools for the other international option (again at the IQO’s discretion).
This plan was designed to yield sufficient evidence of the quality of the data collection
sessions across the surveyed populations.
The IEA Amsterdam received documentation of the international quality control visits to
schools from all participating countries and economies. IQOs successfully conducted
968 school visits in 49 countries/economies,2 with the exception of Australia,
the Netherlands, South Africa and Sweden. Dutch schools teaching at ISCED level 1 took
industrial action during the country’s survey administration window, meaning the observer
could only conduct 19 of the 20 scheduled school visits. In Sweden, the observer
interviewed the required number of SCs (i.e. 20) but had to interview 2 of them by
telephone rather than in person. Similar measures were implemented in Australia, where
the IQO conducted 14 school visits/face-to-face interviews for the ISCED levels 1 and 2
surveys, but 6 telephone interviews for the TALIS-PISA link option. The IQO in
South Africa conducted only 18 school visits due to unforeseen personal circumstances and
practical constraints.
The IEA Amsterdam instructed IQOs to select multiple replacement schools in addition to
the 20 initially selected schools, in case problems occurred (e.g. declines in survey
participation and difficulties arranging school visits) that prevented them from visiting their
originally selected schools. Before beginning their field work, IQOs had to ask the IEA
Amsterdam to approve their respective lists of selected schools and replacement schools.
In general, observers managed to resort to their pre-selected replacement schools when
issues arose with regard to any of the 20 initially selected schools.
Monitoring of the survey administration in the TALIS-PISA link schools received special
attention. The majority of countries/economies conducted the TALIS 2018 project during
the PISA data collection in the second quarter of 2018. The exceptions were Australia and
Ciudad Autónoma de Buenos Aires (Argentina), where administration of the TALIS-PISA
link questionnaires took place in the third quarter of 2018. Both IQOs conducted the
remainder of their visits accordingly.
1. Note by Turkey: The information in this document with reference to “Cyprus” relates to the southern part of the Island. There
is no single authority representing both Turkish and Greek Cypriot people on the Island. Turkey recognises the Turkish Republic
of Northern Cyprus (TRNC). Until a lasting and equitable solution is found within the context of the United Nations, Turkey shall
preserve its position concerning the “Cyprus issue”.
2. Note by all the European Union Member States of the OECD and the European Union: The Republic of Cyprus is recognised
by all members of the United Nations with the exception of Turkey. The information in this document relates to the area under the
effective control of the Government of the Republic of Cyprus.
During their school co-ordinator in-situ interviews, IQOs asked the SCs if the teacher
population being surveyed at the school comprised ISCED level 1, ISCED level 2, ISCED
level 3 or TALIS-PISA link teachers. Table 7.2 compares the expected distribution of
school visits across the surveyed populations against the realised distribution, as reported
by the IQOs. The data show that the realised distribution of school visits deviated slightly
from the planned distribution.
Table 7.2. Comparison of the planned and realised school visit design
School visits ISCED level 1 (%) ISCED level 2 (%) ISCED level 3 (%) TALIS-PISA link (%) Total (%)
The results of the IQOs’ school visits across the different surveyed target populations given
in the following sections are presented together, not separately by TALIS participant.
7.4.1. SC information
In all countries and economies, approximately 96% of SCs were members of the school
staff. Almost half of the co-ordinators were heads of subject or year level and/or another
member of the school management team (42%); about 23% of them were principals and
approximately 19% were teachers. In Chile, for example, co-ordinators were external to
the schools and had other jobs related to the field of education. Across the participating
countries/economies and target populations, the percentage of SCs responsible for
one TALIS school only was 96%.
A large majority of SCs stated that the attitudes of other school staff towards TALIS 2018
were slightly more neutral (about 52%) than positive (fewer than 48%), while less than 1%
of the SCs reported negative attitudes. The negative or indifferent attitudes of the staff
members were commonly characterised as a result of the survey period overlapping with
busy periods in the school year and school staff’s limited knowledge about TALIS.
As documented by the IQOs, 31% of the SCs reported that someone other than the SC
encouraged teachers to join the survey. These people used special instructions, motivational
talks and/or incentives as encouragement. School principals, for example, tended to
encourage participation by delivering motivational or instructional talks. When this
occurred, it was usually during staff meetings or at a group session emphasising the
importance of the survey and of sound survey distribution procedures.
The SCs described teacher participants as broadly co-operative, with between roughly 40%
and 55% of them considering teachers “moderately co-operative or extremely
co-operative” (Table 7.3). Less than 0.4% of the co-ordinators considered teachers “hardly
co-operative at all”. In general, according to roughly three-quarters (76.3%) of the
co-ordinators, teacher respondents felt comfortable answering the survey questions;
approximately 15% of co-ordinators did not know whether teachers felt comfortable.
Table 7.3. IQO-reported interview responses of the SCs to questions about teacher
co-operation and comfort when participating in TALIS
IQOs reported that SCs administered their survey in accordance with the international
procedures. Just over half of the SCs (50.9%) from across the participating countries and
economies adhered to particular data protection rules at the national level in addition to
TALIS 2018 standardised procedures.
Table 7.4. IQO-reported interview responses of the SCs to questions about their experiences,
training and initial survey preparation
Not answered
Questions Yes (%) No (%)
(%)
Did the SC receive a leaflet from the NPM explaining the purpose of TALIS? 91.6 8.4 x
Has the SC previously served as a school co-ordinator for any other survey and/or 36.5 63.2 0.3
assessment (national and/or international)?
Has the SC attended a training session designed for the TALIS 2018 main survey school 43.4 56.5 0.1
co-ordinators?
Did the SC have any difficulty understanding the survey procedures?
Purpose of the survey 5.1 94.9 x
Survey administration (including administration, participation and security/confidentiality 5.3 94.7 x
arrangements)
Survey return procedures (online and paper) 6.4 93.6 x
Did the SC experience any difficulties completing the teacher tracking form or teacher listing 6.8 91.2 2.0
form?
Did the SC experience any inconsistent communication with the NPM, resulting in delays or 2.5 97.4 0.1
unexpected changes?
Did you use the Teacher Tracking Form and Teacher Listing Form? 97.7 2.0 0.2
Is there anyone listed on the Teacher Listing Form who is NOT a teacher? 5.0 95.0 x
NOTE: School staff who do NOT qualify as teachers are out-of-scope for the purposes of
TALIS, and include substitute, emergency or occasional teachers.
x: not applicable.
Note: Percentages were derived from school co-ordinator responses during 976 interviews in total.
In terms of following pre-survey procedures, most countries used the teacher listing form
and the teacher tracking form, as evidenced by the majority of SCs stating that they used
the forms (approximately 98%). Generally, most SCs did not experience difficulties
completing the teacher listing form and teacher tracking form (roughly 91%). The SCs
prepared the teacher listing forms as part of the within-school sampling process, as
described in the School Co-ordinator Manual, which provided detailed information about
who to include on the form.
About 7% of the SCs stated they experienced some difficulty completing the teacher
tracking and teacher listing forms. In some cases, SCs struggled to list eligible teacher
respondents due to teachers working at more than one level of education (international
options) and/or because they did not have access to some of the information requested on
the forms (subjects and names). The IQO data revealed those listed on the teacher listing
form qualified as a teacher in almost all cases (about 95%). In some cases, the lists included
substitute teachers and teachers who were on long-term leave.
The IEA Amsterdam asked IQOs to collect the teacher listing and tracking forms from the
NPMs for every school selected for international quality control in their respective
country/economy and to cross-check the information recorded in the forms against the
forms prepared by the SCs. The majority of SCs (around 96%) confirmed that the IQOs
possessed a complete list of all the teachers employed at each particular school who were
teaching the surveyed population (i.e. ISCED level 1, ISCED level 2, ISCED level 3 and
the TALIS-PISA link). Discrepancies usually corresponded to staff turnover, teacher
absences and maternity leave.
Most SCs (approximately 95%) felt that their respective School Co-ordinator Manual
worked well, with fewer than 6% reporting the need for improvements to the document.
Some TALIS 2018 countries/economies, for example, Austria and England
(United Kingdom), did not produce a manual for SCs but opted to give them letters or other
forms of information, or, in the case of Italy, to run webinars. Some countries/economies
added supplementary material to the manual, such as short videos and screenshots.
Some SCs thought that more information on the survey itself could have been provided in
addition to the procedural information and that the document was overlong and could,
therefore, be shortened. Overall, NPMs were, as reported by the IQOs, highly consistent in
their communication with the SCs, resulting in few delays or unexpected changes.
Approximately 3% of the SCs who answered this question experienced minor
communication issues.
Table 7.5. IQO-reported interview responses of the SCs about explaining TALIS to
participants
Note: Percentages derived from school co-ordinator responses during 976 interviews in total.
Table 7.6. IQO-reported interview responses of the SCs about distribution procedures
Not answered
Questions Yes (%) No (%)
(%)
Did the SC distribute the cover letters (online data collection) 97.4 2.6 x
and/or questionnaires (paper data collection) to participants
in a confidential manner?
Did the SC distribute the cover letters (online data collection) 97.2 2.7 0.1
and/or questionnaires (paper data collection) in accordance
with the teacher tracking form?
x : not applicable.
Note: Percentages derived from school co-ordinator responses during 976 interviews in total.
Table 7.7. IQO-reported interview responses of the SCs about distribution time
Note: Percentages derived from school co-ordinator responses during 976 interviews in total.
With regard to distribution of the survey instruments, approximately 97% of the SCs stated
that they disseminated the materials in a confidential manner. Both the principal
questionnaires and the teacher questionnaires (paper data collection) and/or cover letters
(online data collection) were typically distributed to participants on the same day or within
two to five days after the co-ordinators’ receipt of the instruments.
Similarly, 97.2% of the SCs stated that they distributed the cover letters (online data
collection) and/or questionnaires (paper data collection) in accordance with the teacher
tracking form. The teacher tracking forms and teacher listing forms were often kept in
secure storage, such as the SC’s or the principal’s office. Administering the survey online
was the default mode of data collection; about 88% of the SCs who provided a response
reported that their school administered all the questionnaires online. Bulgaria, Japan, and
Mexico conducted exclusively paper data collection.
The ethics and integrity of the survey relied on IQOs and SCs recognising and respecting
boundaries on information sharing. The manner in which data were collected during TALIS
2018 ensured the anonymity of teachers and principals in the reporting of results. It also
ensured that any information encountered that may have identified the teachers or
principals participating in the survey remained confidential.
When SCs were asked about confidentiality provisions, around 89% of them said they were
the only people to have access to the teacher tracking form and teacher listing form
(Table 7.8). Fewer than 4% of the SCs stated that someone other than themselves had
access to the completed questionnaires (paper data collection), which implies that
confidentiality was primarily upheld across the sampled schools.
Table 7.8. IQO-reported interview responses of the SCs to questions about security and
anonymity
Note: Percentages derived from school co-ordinator responses during 976 interviews in total.
As documented in the school visit records, SCs reported that fewer than 8% of sampled
teachers refused to participate in TALIS. The reasons why non-respondents refused to
participate included lack of time, upcoming retirement, illness, and teachers’ unions or
associations advising against participation.
Table 7.9. IQO-reported interview responses of the SCs to questions about their general
impressions of the survey administration
Not answered
Question Yes (%) No (%)
(%)
Note: Percentages derived from school co-ordinator responses during 976 interviews in total.
IQOs asked SCs to indicate whether any of the TALIS participants approached them to
discuss or ask questions about the survey. Approximately 13% of the SCs said participants
asked them about the purpose of the survey, around 15% said participants asked them to
clarify survey items and about 8% said participants questioned them about the survey return
procedures. Around 2% of SCs said participants asked them about an error they had spotted
and 10% of the SCs reported dealing with other questions about the survey (e.g. queries
about confidentiality or requests for help to access the online questionnaire).
The majority of SCs felt the survey distribution process had gone very well, overall, and
without any problems. Of those interviewed, about 10% reported experiencing some
problems but still deemed the process satisfactory. Suggestions for improvement included
additional time to complete the survey and more detailed instructions on completing the
teacher tracking form.
The SC and IQO feedback implies that teacher respondents remained supportive of TALIS
and understood the importance of the project despite concerns some of them had about the
length of the survey and not having, in their opinion, sufficient time to complete it.
IQOs recorded feedback from the translation verification process and their review of the
verified national instruments and School Co-ordinator Manual(s) in the translation
verification report. The IQOs completed one translation verification report per surveyed
language (maximum of two national languages). The reports provided a starting point for
determining whether any translation problems made the item(s) internationally non-
comparable. Translation verification was implemented to help ensure the international
comparability of the survey instruments.
As an innovative endeavour, the TALIS 2018 Consortium developed three extra questions
designed to test the feasibility of adding quality control questions to the teacher
questionnaire delivered online. The questions were administered to 25 660 teachers
representing a 10% subsample of the teachers participating in TALIS 2018. Of the
three questions developed, two functioned as expected. Since the overall response rate for
these 2 questions was above 95%, the questions could be used for further cycles. The third
question was affected by a misunderstanding and would require further revision for future
administration.
The survey activities questionnaire covered all aspects of survey administration. The ISC
prepared this questionnaire online and delivered it to NPMs for Southern Hemisphere
participants in February 2018, after they had completed data collection. NPMs for Northern
Hemisphere participants received their credentials and information on the survey activities
questionnaire in April 2018. The questionnaire, which consisted of 9 content sections and
156 questions, asked the national study centres to provide information about all survey-
related activities and the extent to which procedures and guidelines had been followed. The
questionnaire also gave NPMs an opportunity to provide valuable feedback about all
aspects of survey administration, including survey procedures and manuals, guidelines,
support materials and software. All TALIS participants completed the questionnaire
between March and December 2018. The following subsections present the results of this
survey.
schools or already existing in schools. During TALIS 2013, 21% of national centres used
other methods; during TALIS 2008, 62% did so.
With regard to data protection and confidentiality laws, 13 countries/economies reported
restrictions on using teacher names on the listing and tracking forms and the national
questionnaires. In these instances, the countries/economies used numbers or codes on the
forms.
Notes
1
Countries administering additional international options (ISCED level 1, ISCED level 3 and the
TALIS-PISA link) were also required to submit the national versions of these instruments.
2
The distinct educational systems in Belgium were taken into account, leading to the recruitment of
one IQO for the Flemish community and one IQO for the French community.
This chapter offers an overview of the approach and strategy used to create the TALIS 2018
international database (IDB). It describes the data-entry and verification tasks carried out
by the national study centres and the exchange of data and documentation between these
centres and the international study centre at the IEA. It also describes the integration of
data from the paper and online administration modes and the data editing and database
creation procedures implemented there, including the detection and resolution of
inconsistencies in the data. The final section of the chapter details the interim data
produced and the steps that all involved centres took to confirm the IDB’s accuracy,
integrity and validity.
8.1. Overview
Creating the TALIS 2018 international database (IDB) and ensuring its integrity required
close co-ordination and co-operation among the international study centre (ISC), Statistics
Canada, the national project managers (NPMs) and the OECD Secretariat.
The primary goals of this work were to ensure that:
all national adaptations to questionnaires were reflected appropriately in the
codebooks and corresponding documentation
all national information eventually conformed to the international data structure
and coding scheme
any errors, such as logical inconsistencies or implausible values given by
respondents or those occurring during data entry, were minimised as much as
possible.
The quality control measures applied throughout the process were identical, to the greatest
extent feasible, for all four target populations: the ISCED level 2 core, ISCED level 1,
ISCED level 3 and the TALIS-PISA link.
The IEA Hamburg supplied the national centres with the IEA Data Management Expert
(DME) software and the TALIS 2018 Main Survey Operations Procedures, Unit 5 (internal
document), which describes the process and rules associated with manual data entry, via
the DME software, of the teacher and principal questionnaires. It also provides information
about the data structure.
The IEA Hamburg also held a three-day data management training session in Rome, Italy,
in October 2016 that covered software use, procedures for national adaptations, and rules
and procedures for data entry. The seminar was specifically targeted at the national team
member(s) responsible for data management and liaising with the IEA Hamburg. In
addition, after administration of the TALIS 2018 field trial in 2017, the TALIS national
project managers (NPMs) received information about the outcomes of the field trial and
resulting software improvements at the third NPM meeting in Lisbon, Portugal, in
July 2017.
This chapter describes the additional steps taken to build the IDB and assure the quality
and accuracy of the TALIS 2018 data.
The default mode of data collection during TALIS 2018 was online questionnaires.
However, paper questionnaires were still important. Of the 48 participating
countries/economies, 37 used the online mode exclusively and 3 used only the paper mode.
The remaining eight countries/economies applied a mixed-mode design, in which they used
the paper mode in addition to the online mode (or vice-versa) for particular ISCED levels,
for either teachers or principals, for selected schools only or for particular school principals
or teachers. National centres had to ensure that individual respondents who refused to
participate in the online mode or did not have access to the Internet received a paper
questionnaire, thereby minimising non-response as a result of a forced administration
mode.
The data from the two different collection modes were later merged into a single set per
ISCED level and country/economy. Potential sources of error originating from the use of
the two parallel modes had to be controlled for and reduced as much as possible to ensure
uniform and comparable conditions across modes and across countries. The design
established several general principles to achieve this:
questionnaires in both modes were self-administered and comparable in terms of
layout and appearance
● the same sample design and procedures were used to identify respondents
● the same methods were used to contact respondents and to validate their
participation
● both modes of data collection occurred over the same period of time.
Notable differences between the two collection modes included approaches to skipping
questions (manually on paper, automatic on line) and the possibility of validating responses
in real time in the online mode.
The electronic versions of the TALIS questionnaires could be completed only on line. No
other options, such as sending/receiving PDF documents by email or printing out the online
questionnaires and mailing them to the national centres, were permissible. Because the
online data collection for TALIS was designed to ensure a standardised educational survey
and to accommodate specific operations, successful administration of the electronic
questionnaires relied on all participating countries/economies adhering to the technical
standard that required them to use the software provided by the TALIS Consortium.
To properly sequence preparation tasks and processes and to ensure comparability of data,
the paper versions of the two questionnaire types (i.e. principal and teacher) had to be
finalised in terms of translation and layout, even if it was likely that all or almost all of the
data would be collected on line. After these final paper versions of the questionnaires had
been converted for the online mode, their structure, text and layout were subject to final
verification.
In addition to these requirements, the design ensured that online respondents needed only
an Internet connection and a standard Internet browser. No additional software, particular
operating system or particular make or version of browsers were required.
The navigational concept for the online questionnaire had to be as similar as possible to
that of the paper questionnaires. Respondents could use “next” and “previous” buttons to
navigate to an adjacent page, actions akin to flipping physical pages. In addition, a
hyperlinked “table of contents” mirrored the experience of opening a specific page or
question of a paper questionnaire. While most respondents followed the sequence of
questions directly, these features allowed respondents to skip or omit questions just as if
they were answering a self-administered paper questionnaire.
To further ensure the similarity of the two sets of questionnaires, responses to the online
questionnaires were not mandatory, evaluated or enforced in detail (e.g. hard validations or
a strict sequence). Instead, some questions used soft validation, such as respondents being
asked to give several percentage numbers that would supposedly add up to 100%. For these
questions, the sum was constantly updated according to the respondent’s entries and was
highlighted in red as long as it differed from 100%. Even if a response remained red,
respondents could proceed to the next question.
Certain differences in the representation of the two modes remained, however. To reduce
response burden and complexity, the online survey automatically skipped questions not
applicable to the respondent, in contrast to the paper questionnaire, which instructed
respondents to proceed to the next applicable question. Rather than presenting multiple
questions per page, the online questionnaire proceeded question by question.
While vertical scrolling was required for a few questions, particularly the longer questions
with multiple “yes/no” or Likert-type items, horizontal scrolling was not. Also, because
respondents could easily use visual cues to estimate the length and burden of the paper
questionnaires, the online questionnaires attempted to offer this feature through progress
counters and a “table of contents” that listed each question and its response status. Multiple-
choice questions were implemented with standard HTML radio buttons. Once respondents
reached the end of the online questionnaires, they were presented with a summary of the
questions they had not answered.
National centres were provided with a tool to monitor online participation. A restricted
version of this tool was also available for school co-ordinators, if requested. While NPMs
could see summary information (e.g. first login, last login, total of logins, progress of
answers) and raw data for all principals and teachers, school co-ordinators could only view
the type of summary information for their school that would prompt them to initiate
appropriate follow-up activities. They could not access actual responses.
Each national centre was responsible for transcribing the information from the principal
and teacher questionnaires into computer data files. National centres entered responses
from the paper questionnaires into data files created from an internationally predefined
codebook. This contained information about the names, lengths, labels, valid ranges (for
continuous measures or counts) or valid values (for nominal or ordinal questions) and
missing codes for each variable in each of the two questionnaire types. Before data entry
commenced, national data managers (NDMs) were required to verify the nationally adapted
codebook structure that reflected all ISC-approved adaptations (e.g. a nationally added
response category) made to the national questionnaire versions. These adapted codebooks
then served as templates for creating the corresponding data set.
In general, the ISC instructed national centres to discard any questionnaire that was unused
or returned completely empty and to enter any questionnaire that contained at least one
valid response. To ensure consistency across participating countries, the basic rule for data
entry in DME required national staff to enter data “as is”, without any interpretation,
correction, truncation, imputation or cleaning. Any inconsistencies that remained after this
data-entry stage were dealt with at the time of data cleaning (see below).
The rules for data entry meant that:
Responses to categorical questions were generally coded as “1” if the first option
(checkbox) was used, “2” if the second option was marked, and so on.
Responses to “check-all-that-apply” questions were coded as either “1” (marked)
or “9” (omitted or invalid).
Responses to numerical or scale questions (e.g. school enrolment) were entered “as
is”, that is, without any correction or truncation, even if the value was outside the
originally expected range (e.g. if a teacher reported that he or she spent 80 hours a
week teaching students in school). If countries needed to enter values that exceeded
the defined variable width, they entered these few values on an Excel sheet and
submitted it to the ISC, which then included these values during its data processing
work.
of the layout verification process, to electronic copies of the national versions of all
questionnaires and the final national adaptation forms (NAFs).
8.4. Data checking, editing and quality control at the IEA Hamburg
Once the national centres submitted their data to the ISC, data processing commenced. The
objective of this process was to ensure that the data adhered to international formats, that
information from principals and teachers could be linked across different survey files and
that the data accurately and consistently reflected the information collected within each
participating country/economy. The ISC went to great lengths to ensure that the data
received from participating countries/economies were internationally comparable and of
high quality. The foundation for quality assurance had been laid down before the data were
submitted to the ISC through the provision of manuals, training and software designed to
standardise a range of operational and data-related tasks, and through verification of the
content and layout of the NAFs, paper questionnaires, online questionnaires and
codebooks.
The WinW3S software performed the within-school sampling operations, strictly
adhering to the sampling rules defined by TALIS (see Chapter 5 for more details
on school sampling). The software created all necessary listing and tracking forms
and stored school-specific and teacher-specific information, such as gender and
participation status. The software also generated login credentials that consisted of
the unique ID and the corresponding checksum (the five-digit validation code) that
were used for both survey administration modes. For the purpose of paper
administration, WinW3S created questionnaire labels that included the generated
login credentials. It also included a participation rate estimator that kept NDMs up
to date on their current participation rate according to the already returned paper
questionnaires or submitted online questionnaires. The ISC asked NDMs to begin
uploading their weekly participation rate reports as soon as data collection began.
The DME software enabled entry of all questionnaire data in a standard,
internationally defined format. Data entered with the DME software were
automatically validated. This process included validation of login credentials and a
range of other issues, such as the uniqueness of the ID or out-of-range or otherwise
invalid codes. Whenever the software flagged such issues, it also prompted the
individuals entering the data to resolve or to confirm the inconsistencies before
resuming data entry. In addition, special variables called “check variables” were
used during data entry to avoid any individual entering values for the wrong
question. The ISC asked national study centres to use the DME software for data
entry. However, if the centres’ countries/economies wanted to use different tools to
enter their data, the DME gave them the option of importing these data and
verifying them with the same range of checks used by those participating
countries/economies entering the data directly via the DME software. The software
also included a range of data-verification checks that NDMs had to perform during
data capture and after data entry.
A complex study such as TALIS 2018 required a correspondingly complex data cleaning
design. Accordingly, the ISC developed processing tools in Structured Query Language
(SQL) and, where necessary, in Statistical Package for the Social Sciences (SPSS). The
ISC took the following steps to ensure that programmes ran in the correct sequence, that
no special requirements were overlooked and that the cleaning process was implemented
independently of the persons in charge.
Before the data-cleaning programmes were applied to real data, all of them were
thoroughly tested using simulated data sets containing all expected problems or
inconsistencies.
To document versions and updates, all incoming data and documents were
registered in a specific material receipt database. The date of arrival and any
specific issues meriting attention were recorded.
All national adaptations and all detected deviations from the international data
structure were recorded in a “national adaptation database” and verified against the
national adaptation form (NAF), the national instruments, the codebooks and the
contents of the data. The reports from this process are available for data analysts in
the TALIS 2018 and TALIS Starting Strong Survey 2018 User Guide (OECD,
forthcoming[1]).
Cleaning was organised according to rules strictly and consistently applied to all
national data sets, making deviations in the cleaning sequence impossible.
All systematic or manual corrections made to data files were implemented and
recorded in specific cleaning reports that the TALIS Consortium and the NPMs
then reviewed and approved.
On completion of the data cleaning for a participating country/economy, all
cleaning checks were repeated from the beginning to detect any problems that
might have been inadvertently introduced during the cleaning process itself.
Figure 8.1 provides a schematic overview of this iterative process conducted in
co-operation with the national centres. The following subsections of this chapter give a
more detailed description of the sequential data-cleaning steps displayed in Figure 8.1.
The structure check implemented at the ISC looked for differences between the
international and the national file structures. As described above, some participating
countries/economies made structural adaptations to the questionnaires, the extent and
nature of which differed greatly across the countries/economies. Whereas some
participating countries/economies administered the questionnaires without any change,
except for translations and necessary cultural adaptations, others inserted or removed
questions or options within existing international variables or added entirely new national
variables.
Given the associated risk of deviating from the international data structure, NPMs wishing
to make such changes had to follow certain strict rules to allow unequivocal integration of
nationally adapted variables for international comparison. Where necessary, the ISC
modified the data according to the international design to ensure that the resulting data were
internationally comparable. For instance, the ISC recoded (mapped) additional national
options in multiple-choice questions in a way that ensured they adhered to the international
code scheme. National variables were created to hold the original values for later use in the
national reports.
NPMs and NDMs received detailed reports on structural deviations together with
documentation on how the ISC resolved them. In a few cases, data were not available for
certain variables because the corresponding question was not administered nationally; see,
in this regard, the TALIS 2018 and TALIS Starting Strong Survey 2018 User Guide (OECD,
forthcoming[1]). In a few instances, data had to be removed from the IDB because the
information was not internationally comparable due to errors in translations that were
spotted only after the data had been collected.
The individual responses to percentage questions were summed. These values were set to
“omitted or invalid” if they fell outside the 90 to 110 range (PQ-11 and PQ-21), or the 45
to 110 range (TQ-39), or if any of them were larger than 100.
Variables with implausible numerical values were also set to “omitted or invalid”. For
example, TQ-38 asked respondents to give the average number of students in the target
class. Values exceeding 100 were set to “omitted or invalid”.
Finally, inconsistencies between the listing information in the teacher tracking forms and
the actual responses of the teachers regarding their age and gender were resolved in a
manner that gave precedence to the teacher-supplied information.
The number of inconsistent or implausible responses in the data files varied from
one country/economy to another, but no national data were completely free of inconsistent
responses. Each problem was recorded in a database and identified by a unique problem
number. The entry also included a description of the problem and of the automatic action
taken by the software programme or the manual action taken by ISC staff. Staff referred
issues that could not be corrected using systematic rules to the relevant NPM so that he or
she could check the original data collection instruments and tracking forms and trace the
source of the inconsistency. Whenever possible, staff at the ISC suggested a solution and
asked the NPMs either to accept it or to propose an alternative. Data files were then updated
to reflect the agreed-upon solutions. Both systematic corrections and those apparent on a
case-by-case basis were applied directly in SQL programme syntax and carried out
automatically for each cleaning run.
If an NPM could not solve a problem by inspecting the instruments and forms or could not
suggest a satisfying solution or explanation, the TALIS Consortium defined the final
cleaning rules. The ISC and the OECD Secretariat together agreed to any systematic
content edits and documented them for use by the NPM.
The IEA Hamburg and Statistics Canada agreed to a special treatment for TQ-12/13 for
weighting purposes and calculation of the teacher multiplicity factor (WGTADJ4) (see
Chapter 9 for more details).
If TQ-12 is “yes” (1) and TQ-13 is “omitted or invalid” or zero (0) → recode TQ-12
to “no” (2) and TQ-13 to “logically not applicable”.
If TQ-12 = “no” (2) and TQ-13 is zero (0) or one (1) → recode TQ-13 to “logically
not applicable”.
If TQ-12 is “no” (2) but TQ-13 is two (2) or more → recode TQ-12 to “yes” (1).
8.4.5. Final action for yes/no lists with more than two items
For those questions with lists that were partially answered with “yes” and “omitted or
invalid”, all omitted responses were recoded to “no”. These questions included PQ-07, PQ-
15 (Part A), PQ-33, PQ-38, PQ-39, TQ-06A (Part A), TQ-20, TQ-22, TQ-23, TQ-24, TQ-
26, TQ-31, TQ-47, TQ-56 and TQ-57.
8.4.6. Final action for check-all-that-apply questions with more than two items
For those questions that had checkboxes that were partially “checked”, “not checked” and
“omitted or invalid”, all omitted responses were recoded to “not checked”; for checkboxes
that were all marked “not checked”, all responses were recoded to “omitted or invalid”.
The relevant questions were PQ-20, PQ-24, TQ-15, TQ-19 and TQ-29.
Treatment for PQ-16: if value is zero (0) in the questionnaire → set to “omitted or
invalid”.
Treatment for PQ-42: if sum of self-reported principal age (PQ-2) and PQ-42 is
higher than 100 → set to “omitted or invalid”.
Treatment for TQ-05: if value is between 40 and 99 → add 1900 so values are
changed to 1940–1999.
Treatment for TQ-05: if value is not between 1940 and 2018 → set to “omitted or
invalid”.
Treatment for TQ-11a/b: if value is higher than 58 → set to “omitted or invalid”.
Treatment for TQ-11c/d: if value is higher than 57 → set to “omitted or invalid”.
Treatment for TQ-16/17/18: if value is higher than 120 → set to “omitted or
invalid”.
Treatment for TQ-38: if enrolment is zero (0) or larger than 100 → set to “omitted
or invalid”.
Treatment for TQ-50: if sum of self-reported teacher age (TQ-2) and TQ-50 is
higher than 100 → set to “omitted or invalid”.
8.4.9. Final recoding for inconsistent teacher age and gender in listing and
questionnaire information
The recoding instructions for PQ-1, TQ-1, TQ-2, GENDER (gender on listing form) and
ITBIRTHY (birth year on listing form) were as follows:
Teacher birth year (ITBIRTHY): if value is outside the range of 1941–2000 → set
to “omitted or invalid”.
Gender (TQ-1 vs. GENDER): (a) believe questionnaire information and substitute
listing information gender in case it is missing or inconsistent; (b) impute missing
questionnaire value from listing if questionnaire variable was omitted.
Teacher age (TQ-2, ITBIRTHY): (a) believe questionnaire information and delete
listing information if inconsistent; (b) impute missing questionnaire value from
listing form.
Teacher age (TQ-2): if outside of range of 18-76 → set to “omitted or invalid”.
Principal age (PQ-2): if outside of range of 23-73 → set to “omitted or invalid”.
Building the TALIS international database was an iterative process during which the ISC
provided the OECD Secretariat and NPMs with a new version of data files whenever a
major step in data processing was completed. This process guaranteed that NPMs had a
chance to review their data and run additional plausibility and statistical checks to validate
the data. The data products that the ISC released to the OECD Secretariat and each NPM
included the teacher and principal data files as well as data summaries. All interim data
were made available to the OECD Secretariat in full, whereas each participating
country/economy received only its own data.
The ISC sent the first version of cleaned and weighted data to the OECD Secretariat at the
end of September 2018, by which time all known identification, linkage and content issues
in these data had been resolved. Estimation weights and variables facilitating variance
estimation were also included. To protect respondents’ identity, the ISC scrambled the
respondents’ IDs. These data were also used to produce the first set of draft tables for the
international report, presented to NPMs at the fourth NPM meeting in Seoul, Korea, in
October 2018. Before this meeting, all NPMs received a version of their own cleaned and
weighted data, giving them a chance to review their data and the tables produced by the
OECD Secretariat for accuracy and validity.
During the fourth NPM meeting, NPMs were able to raise any issues concerning their data
that had, thus far, gone unnoticed. This process resulted in a second, updated data version
that concluded the main survey’s field work and included scale scores. The ISC sent this
version to the OECD Secretariat and NPMs at the end of November 2018.
All interim data products were accompanied by detailed data processing and weighting
documentation and summary statistics. The latter contained weighted and unweighted
univariate statistics and frequencies for all questionnaire variables for each
country/economy. For categorical variables, which represent the majority of variables in
TALIS, the percentages of respondents choosing each of the response options were
displayed. For numeric or count variables, various descriptive measures were reported.
These included the minimum, the maximum, the mean, the standard deviation, the median,
the mode, percentiles and quartiles. For both types of variables, the percentages of missing
information due to respondents omitting or not reaching a particular question were
reported. These summaries were used for a more in-depth review of the data at the
international and national levels in terms of plausibility, unexpected response patterns,
suspicious profiles and so on.
For the draft and final IDB, data cleaning at the ISC ensured that the information coded in
each variable was, in fact, internationally comparable, that national adaptations were
reflected appropriately in all variables concerned and that all records could be successfully
linked across the two levels. For countries/economies participating in the TALIS-PISA link
option, a variable was included that enabled later linkage to schools in the PISA 2018
database (PISASCHOOLID).
The interim data products described above and the draft and final (public-use) international
databases had two key differences:
All interim products included one record for each sampled unit (school or teacher),
even if the school did not return the corresponding questionnaire or returned it
empty. In contrast, the draft and final IDB included only records that satisfied the
sampling standards. Data from those units that either did not participate or did not
pass adjudication (e.g. because within-school participation was insufficient) were
removed.
To protect the confidentiality of respondents’ information, disclosure avoidance
measures were applied at the international level (1) consistently for all participating
countries/economies and (2) in specific national data sets. These measures were
implemented for all data versions and exports of the IDB for use by all other
countries/economies and public users.
The measures applied to all international-level data sets involved the following:
Scrambling of the teacher identifier (IDTEACH) and school identifier
(IDSCHOOL): Because these identifiers were scrambled, they did not match those
used during data collection; however, the structural link between the school and
teacher level (the variable IDSCHOOL in the teacher file and the first four digits
of any IDTEACH) was maintained. Unique matching tables were created for each
country/economy and made available to authorised individuals within that
country/economy.
Variables used purely for the stratification of the teacher sample, that is, birth year
(ITBIRTHY) and gender (GENDER): These were removed. Only the gender
(TT3G01) and age (TT3G02) variables, as collected in the questionnaire, were
retained.
Variables used purely for stratification of schools, that is IDSTRATE and
IDSTRATI: These were removed to prevent identification of geographical or
organisational groups. Because the stratum information is mostly of interest for
national-level analysis, it was of course made available to the country/economy
concerned. Researchers from other countries wanting to conduct analysis by
stratification will need to request the stratification variables directly from the
respective country/economy.
Information used in the calculation of final sample and replicate weights (for the
school level, WGTFAC1 and WGTADJ1; for the teacher level, WGTFAC1,
WGTADJ1, WGTFAC2, WGTADJ2, WGTADJ3 and WGTADJ4): This
information was removed from the IDB because it could allow identification of
stratification cells.
Replication zone and unit variables (BRRSZONE, BRRSREP, BRRTZONE and
BRRTREP): These were dropped from public-use micro-data because they could
enable indirect identification of schools.
The process of building the IDB complied, at all times, with the OECD’s rules for the
processing and transfer of personal data and, where relevant, with rules stipulated by
national or regional legislation for the protection of personal data (e.g. the European
Union’s General Data Protection Regulation).
To protect its respondents’ privacy, Iceland decided to withdraw all data from the IDB.
This information can, therefore, be procured only by applying directly to this country.1
After each NPM and the OECD Secretariat had agreed on data-release policy and
confidentiality agreements, a draft IDB that included data from all participating
countries/economies was made available at the end of February 2019, prior to publication
of the first international report in June 2019. This release enabled participating
countries/economies to replicate results presented in the draft chapters of the international
report. This data version was also used in an international database training session that
ISC staff held in Lüneburg, Germany, in March 2019. However, only accredited members,
that is, the people who signed the confidentiality agreement with the OECD Secretariat,
had access to it. Non-accredited members had access to a separate mock database.
The final public-use IDB was scheduled for release in May 2019, supplemented by full
documentation in the TALIS 2018 and TALIS Starting Strong Survey 2018 User Guide
(OECD, forthcoming[1]). The database, which contains data from schools and teachers from
83 different samples in 48 participating countries/economies across 5 continents, provides
a unique resource for policymakers and analysts.
Although data for all participating TALIS 2018 entities are included in the IDB, the sample
adjudication process determined that the ISCED level 1 teacher and principal data for
Australia and the Netherlands and the ISCED level 2 principal data for Australia cannot
reliably represent the population from which the sample was drawn. The sampling
adjudication variable INTAL18 was therefore set to zero. When conducting any analyses,
database users need to ensure that they use only those cases where INTAL18 equals 1.
Reference
OECD (forthcoming), TALIS 2018 and TALIS Starting Strong Survey 2018 User Guide, [1]
OECD, Paris.
Note
1
Please contact the TALIS team at the OECD to be put in contact with Iceland.
This chapter covers three important aspects of the quality of the TALIS 2018 outcomes. The
first is the weighting of the data to correct for unequal probabilities of selection and to
produce unbiased estimates. Descriptions are provided of how each component of the final
estimation weights was defined and how those components were assembled into the final
estimation weights. The second aspect, participation rates, is also described. Finally, the
balanced repeated replication (BRR) weights for the estimation of the sampling error, the
third aspect, are detailed.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
9.1. Overview
This chapter covers three important aspects of the quality of the TALIS 2018 outcomes:
the weighting of the data to correct for unequal probabilities of selection and produce
unbiased estimates, the participation rates and the estimation of sampling error.
Although the international sampling design was prepared as a self-weighting sampling plan
of teachers (whereby each individual ultimately had the same final estimation weight), the
conditions in the field, school and teacher non-response, and the co-ordination of multiple
samples made it impossible to fulfil that ideal plan. In the end, in most participating
countries, the national sampling plan was a stratified multi-stage probability sampling plan
with unequal probabilities of selection.
Because the sample of schools and principals was an intermediary step, that is, a by-product
of the teacher sampling design, schools and principals had the same design weights. In a
few participating countries, namely Cyprus,1 Iceland, Malta, the Netherlands, the Russian
Federation, Singapore and the United Arab Emirates, the canonical plan was modified to
suit local conditions (see Chapter 5, Annex E). How each component of the final estimation
weight was defined and how those components were assembled into the final estimation
weight are detailed below.
The section of this chapter covering the second aspect presents a description of the
participation rates and how they were computed. Annex E provides the results for each
participating country and each survey in which they participated.
Because of the unequal weights and because of the structure of the samples, sampling error
must be estimated using the design and weights. Failure to do this can translate into severely
biased estimates of sampling error. Correctly estimating sampling error for complex
samples is often a daunting task but simple and approximately unbiased methods are
available. TALIS 2018, like its predecessors, opted for balanced repeated replication (BRR)
not only because of this method’s statistical properties (consistency, asymptotic
unbiasedness) and its portability (one formula fits all types of parameter estimates) but also
because it is comparatively easy to compute.
The last section of this chapter explains how the replicates were created and how the BRR
estimates of sampling error were computed. These estimates of the sampling error were
another key element of the statistical quality of the TALIS survey outcomes.
A more detailed description of the survey design and its implementation can be found in
Chapter 2, 5 and 6 of this report, in the TALIS 2018 Survey Operations Procedures Unit 1:
Sampling Schools, in the TALIS 2018 National Project Managers Manual and in the TALIS
2018 School Co-Ordinator Manual (internal documents).
The statistics produced for TALIS 2018 were derived from data obtained through samples
of schools, school principals and teachers. For these statistics to be meaningful for a
country/economy, they needed to reflect the whole population from which they were drawn
and not merely the sample used to collect them. The process of going from the sample data
to information about the parent population is called estimation. When the sample is
equiprobable, unstratified and unclustered, simple sample averages may suffice as
estimates of population averages (e.g. the average number of ISCED level 2 teachers per
school). However, sample counts do not suffice as estimates of population totals (e.g. the
total number of ISCED level 2 teachers in a country).
The estimation weight or final weight is the device that allows the production of country-
level estimates from the observed sample data. The estimation weight indicates the number
of units that a sampled unit represents. The final weight is the combination of many factors
reflecting the probabilities of selection at the various stages of sampling and the response
obtained at each stage. Other factors may also come into play as dictated by special
conditions to maintain unbiasedness of the estimates (e.g. adjustment for teachers working
in more than one school).
Because TALIS 2018 consisted of a compulsory core segment (ISCED level 2) and
three optional segments (ISCED level 1, ISCED level 3 and the TALIS-PISA link),
estimation weights had to be computed independently for each segment. This requirement
held true even when samples were co-ordinated across TALIS segments (ISCED levels 2
and 3, for example) or across survey programmes (TALIS and TIMSS,2 for example).
Basically, final weights are the product of a design or base weight and of one or many
adjustment factors; the former is the inverse of the selection probability and the latter
compensates for random non-response and other random occurrences that could, if not
accounted for, induce biases in the estimates. These design weights and adjustment factors
are specific to each stage of the sample design and to each explicit stratum used by the
design. Clearly, in instances where the participating countries/economies adapted the
general sample design of TALIS 2018 to their own conditions, the estimation weights had
to conform to these national adaptations.
The following are the conventional notations used in this chapter. As usual, the letters h, i,
and j are used as subscripts, the lower-case letters k, l, m, n, r and t refer to the sample, and
the upper-case letters H, L, M and N refer to the population.
Each participating country has H explicit strata and the index h = 1, …, H points to
the explicit stratum. If no explicit strata were defined, then H = 1.
In each explicit stratum, a sample of size nh schools was drawn from the Nh schools
forming stratum h. The index i = 1, …, nh, therefore, points to the ith sampled
school in stratum h.
Each school i = 1, …, nh within the explicit stratum h has a measure of size (MOS)
noted as Mhi; the sum of the individual measures of size is noted as Mh.
In each responding school, a sample of mhi teachers was drawn from the listing of
Lhi teachers. If no changes had occurred in the school since the creation of the
sampling frame, then Lhi = Mhi, but this was seldom the case. If the selected school
was large enough, mhi = 20 by design. The index j = 1, …, mhi points to the teachers
and mhi can, therefore, differ from 20 if local conditions dictated that the sample
size should differ. For example, if the size of the listing was Lhi = 18, then all
teachers were selected and mhi = 18.
weight is needed to represent this first stage of sampling. If a census sample of schools was
implemented in a country or in an explicit stratum of a country, then the school base weight
is set to 1.
Use of the above notation established the school base weight for each school i = 1, …, nh
and each explicit stratum h = 1, …, H, as:
𝑀ℎ
𝑊𝐺𝑇𝐹𝐴𝐶1ℎ𝑖 = 𝑛 .
ℎ ×𝑀ℎ𝑖
In those countries (Cyprus,3 Iceland, Malta and the United Arab Emirates) where all schools
were selected (i.e. n = N), there was only one stratum, and WGTFAC1i = 1 for all i = 1, ...,
N. In the Russian Federation, where geographical regions were first selected at random, the
weight component corresponding to that selection was incorporated into the school base
weight (WGTFAC1).
9.3.3. Final (estimation) school weight for school and principal data
As described earlier, because the school estimation weight is the product of the school base
weight and the school non-response adjustment factor, it should be used for estimation of
school-related parameters.
The final school weight (school estimation weight) for each participating school i = 1, …,
rh and each explicit stratum h = 1, …, H was given as:
𝑀ℎ 𝑛ℎ −𝑑ℎ
𝑆𝐶𝐻𝑊𝐺𝑇ℎ𝑖 = 𝑊𝐺𝑇𝐹𝐴𝐶1ℎ𝑖 × 𝑊𝐺𝑇𝐴𝐷𝐽1ℎ = 𝑛 × .
ℎ ×𝑀ℎ𝑖 𝑟ℎ
For those countries where a census was conducted, the final school weight was
𝑆𝐶𝐻𝑊𝐺𝑇𝑖 = 1(𝑁 − 𝑑) / 𝑟 = (𝑁 − 𝑑) / 𝑟.
condition held whether or not a principal returned his or her questionnaire. Therefore, some
schools for which principal data were available could be deemed non-participant because
fewer than 50% of the selected teachers returned their questionnaires. An adjustment factor
to the school design weight was thus necessary and it was possible that it would differ from
the adjustment factor computed for the school database.
For each explicit stratum h = 1, …, H, if rh schools participated (i.e. at least 50% of selected
teachers returned their questionnaire) in TALIS 2018 out of the nh selected schools, and if
dh schools were closed or out of scope, the non-response adjustment factor was:
𝑊𝐺𝑇𝐴𝐷𝐽1ℎ =
𝑛ℎ −𝑑ℎ
𝑟ℎ
for participating schools (at least 50% teacher questionnaires)
{ 1 for closed or out-of-scope schools .
0 for non-participating schools (less than 50% teacher questionnaire)
The weight adjustment factor for incidental exclusion of teachers in each participating
school i=1, …, rh in explicit stratum h = 1, …, H, was:
𝐿
𝑊𝐺𝑇𝐹𝐴𝐶3ℎ𝑖𝑗 = 𝐿−ℎ𝑖 .
ℎ𝑖
In this adjustment factor, the numerator is the full school measure of size as listed and the
denominator is the size of the reduced list from which the sample was actually selected.
In the simplest of cases, the sampling design prepared for TALIS 2018 would have yielded
equal weights for all teachers. If we assume (1) that the sample size of schools was
distributed among the explicit strata proportionally to the number of teachers in each
stratum, (2) that samples of 20 teachers could be selected from every selected school,
(3) that the school listings were equal to the measures of size used to select the schools,
(4) that the school listings contained only in-scope teachers, (5) that no incidental exclusion
occurred, (6) that each selected school and teacher participated, and (7) that each teacher
was teaching in only one school, then the final teacher weight will be the same for all the
teachers in the sample:
𝑇𝐶𝐻𝑊𝐺𝑇ℎ𝑖𝑗
= {𝑊𝐺𝑇𝐹𝐴𝐶1ℎ𝑖 × 𝑊𝐺𝑇𝐴𝐷𝐽1ℎ𝑖 }
× {𝑊𝐺𝑇𝐹𝐴𝐶2ℎ𝑖𝑗 × 𝑊𝐺𝑇𝐴𝐷𝐽2ℎ𝑖𝑗 × 𝑊𝐺𝑇𝐴𝐷𝐽3ℎ𝑖𝑗 × 𝑊𝐺𝑇𝐴𝐷𝐽4ℎ𝑖𝑗 }
𝑀ℎ 𝑀ℎ𝑖
={ × 1} × { × 1 × 1 × 1}
𝑛ℎ × 𝑀ℎ𝑖 20
𝑀 𝑀ℎ 𝑀
={ }×{ }=
200 × 𝑀ℎ 20 4000
for h = 1, …, H; i= 1, …, 𝑛ℎ ; j = 1, …, mhi, and where M is the total number of teachers in
the population of interest.
As described in Chapter 5, the sample of schools for the TALIS-PISA link was a subset of
the sample of schools selected to take part in PISA 2018. Given the sequencing of events
between TALIS and PISA, the sampling team could not limit subsampling for the TALIS-
PISA link to schools that had participated in PISA. They, therefore, had to draw the
subsample from the full sample of schools prior to the PISA data collection. However,
because data collection for the TALIS-PISA link was scheduled to take place after
completion of the data collection for PISA (at least, for any given school), the school base
weight was that of the PISA 2012 design, adjusted for subsampling. Where school
non-response occurred, computation of the non-response adjustment was similar to the
process described above. Again, the (TALIS-PISA link) school estimation weight was the
product of the (TALIS-PISA link) school base weight and the (TALIS-PISA link) school
non-response adjustment factor. It should also be used for estimation of the TALIS-PISA
school-related parameters.
Because teacher sampling for the TALIS-PISA link followed the same rules as for the
ISCED levels, the construction of the “PISA teacher” weight followed the same steps: base
weight within a TALIS-PISA school, non-response adjustment within the school, and
multiplicity and exclusion adjustments. The final TALIS-PISA-teacher weight (estimation
weight) was thus the product of the teacher base weight, the three adjustment factors
associated with each participating teacher and the final TALIS-PISA link school weight.
All estimates pertaining to the populations of PISA teachers, therefore, needed to use the
(TALIS-PISA link) final teacher weight.
The quality requirements for TALIS 2013 translated into participation rates (response rates)
for schools and for teachers (see Chapter 10 for individual participants’ results). Reaching
the required levels of participation does not preclude some degree of error in the results but
should reduce reliance on the “missing at random” assumptions made for the non-response
weighting adjustments. Experience and knowledge gained from TALIS 2008 and TALIS
2013 showed that the targets set for TALIS 2018 participation were realistic.
where rh, nh and dh are as defined above. This formula represents the crude proportion of
schools from which a principal questionnaire was received (the unweighted participation
rate is sometimes interpreted as a crude measure of the effectiveness of collection).
The weighted school participation rate was computed as:
𝑟ℎ
∑𝐻
ℎ=1 ∑𝑖=1 𝑊𝐺𝑇𝐹𝐴𝐶1ℎ𝑖
𝑊𝑇𝐷𝑆𝐶𝑃𝐴𝑅𝑇 = 𝑟ℎ .
∑𝐻
ℎ=1 ∑𝑖=1 𝑆𝐶𝐻𝑊𝐺𝑇ℎ𝑖
This formula represents the proportion of the population of schools or principals accounted
for by the participating schools.
Both rates were computed once over the complete set of participating schools (after
replacement) and once over the subset of participating schools in the original selection
(before replacement).
This formula gives the crude ratio of the number of responding teachers in participating
schools with respect to the expected sample size from the participating schools. The
weighted teacher participation rate was, therefore:
𝑊𝑇𝐷𝑇𝑃𝐴𝑅𝑇
𝑟ℎ 𝑡ℎ𝑖
∑𝐻
ℎ=1 ∑𝑖=1 ∑𝑗=1{𝑊𝐺𝑇𝐹𝐴𝐶1ℎ𝑖 × (𝑊𝐺𝑇𝐹𝐴𝐶2ℎ𝑖𝑗 × 𝑊𝐺𝑇𝐴𝐷𝐽3ℎ𝑖𝑗 × 𝑊𝐺𝑇𝐴𝐷𝐽4ℎ𝑖𝑗 )}
= 𝑟ℎ 𝑡ℎ𝑖 .
∑𝐻
ℎ=1 ∑𝑖=1 ∑𝑗=1{𝑇𝐶𝐻𝑊𝐺𝑇ℎ𝑖𝑗 }
Estimation, especially estimation of sampling error for surveys with complex designs such
as TALIS, requires special attention. Both the survey design and the unequal weights are
needed to obtain (approximately) unbiased estimates of sampling error. Not taking this
approach can lead to severe underestimation of the sampling error. While exact formulae
exist in theory for stratified PPS sample designs, the required computations become
practically impossible as soon as the number of primary units selected per stratum exceeds
two.
Over the years, various statisticians have proposed approximate solutions for this problem.
An important class of solutions is that of resampling or replication. The best-known
examples of replication methods are interpenetrating subsamples (Mahalanobis), Balanced
Half-Samples or Balanced Repeated Replication (McCarthy, Fay), the Jackknife
(Quenouille, Tukey, Durbin, Frankel), and the Bootstrap (Efron). For reviews of these
methods, see, for example, Lohr (1999[1]), Rust and Rao (1996[2]) or Wolter (2007[3]).
In a similar vein to what was done for PISA (see, for example, OECD (2009[4])), TALIS
adopted the Balanced Repeated Replication (BRR) for estimation of the sampling error of
the estimates. BRR is a replication method suited to sample designs where exactly
two primary sampling units (PSUs) are selected in each stratum.
The principle of BRR is as follows: each of the two PSUs can provide an unbiased estimate
of the total (or another parameter of interest) of its stratum. If the sampling design
comprises H strata, there are then 2H possible unbiased estimates of the parameter of
interest, obtained by combining either PSU from each of the H strata. The sampling error
of the estimate of the parameter of interest can be directly computed by comparing each of
the 2H estimates with their mean, as is usually done in simple basic statistics. Even with
moderate values of H, the number of unbiased estimates may be quite large (e.g. 25 = 32,
210 = 1 024, 220 = 1 048 576). BRR provides a way to extract from the complete set of 2 H
possible replicates a much smaller subset that gives the same measure of sampling error as
the full set.
Table 9.1. Example of BRR-ready sample design and random assignment of pseudo PSUs
Explicit stratum School ID Zone = pseudo stratum Pseudo PSU Other variables of interest…
1 1001 1 1 … …
1 1002 1 2
1 1003 2 1
1 1004 2 2
2 1005 3 2
2 1006 3 1
2 1007 4 1
2 1008 4 2
… …
Explicit stratum School ID Zone = pseudo stratum Pseudo PSU Other variables of interest…
H … G-1 2
H … G 2
H … G 1
As with the Jackknife Repeated Replication, one of the two pseudo PSUs is dropped, which
means the weight of the remaining PSU doubles. This PSU is the one that is used to
compute an estimate of the parameter of interest. Rather than randomising which PSU to
drop, we used a special matrix (of order 4t) of +1’s and -1’s (the so-called Hadamard matrix)
to indicate which PSU should be kept (+1) and which should be dropped (-1) from each
pseudo stratum in BRR, a process that associated the +1’s with the PSUs numbered 1 and
the -1’s with the PSUs numbered 2. As an example, the Hadamard matrix of order 8 can be
written as:
+1 +1 +1 −1 +1 −1 −1 −1
−1 +1 +1 +1 −1 +1 −1 −1
−1 −1 +1 +1 +1 −1 +1 −1
+1 −1 −1 +1 +1 +1 −1 −1
𝐻𝑎𝑑𝑎𝑚𝑎𝑟𝑑8 = .
−1 +1 −1 −1 +1 +1 +1 −1
+1 −1 +1 −1 −1 +1 +1 −1
+1 +1 −1 −1 −1 −1 +1 −1
(−1 −1 −1 −1 −1 −1 −1 −1)
In this matrix, each column is a BRR replicate and each line is a pseudo stratum or zone.
The matrix entry indicates which pseudo PSU should be kept from each pseudo stratum to
create the BRR replicate. For example, the previous matrix translates into:
For TALIS 2018, and in keeping with what was done in the previous cycles of TALIS and
PISA, a variation of the BRR developed by Fay (1989[5]) was implemented. Instead of a
PSU being completely dropped and the weight of the other one doubled, the final weight
(𝑇𝐶𝐻𝑊𝐺𝑇ℎ𝑖𝑗 ) of every teacher in the PSU indicated by the Hadamard matrix is multiplied
by the replicate factor of 1.5 to get the replicate weight. Likewise, the final weight
(𝑇𝐶𝐻𝑊𝐺𝑇ℎ𝑖𝑗 ) of the teachers in the remaining PSU is multiplied by the replicate factor of
0.5 to get the replicate weight. This strategy removes the risk of completely deleting a
domain.
In cases where there was an odd number of PSUs in an explicit stratum, the last three PSUs
were treated as a zone in the following manner: one of the PSUs was randomly designated
as “+1” while the remaining two were both designated as “-1”. For each replicate, as
indicated by the Hadamard matrix, the weight of the selected unit was multiplied by 1.7071
if it was the single unit and the weights of the remaining pair were multiplied by 0.6464. If
the matrix indicated that the pair should be selected, then the weights of the paired units
were multiplied by 1.3536 and the weight of the single unit was multiplied by 0.2929. This
strategy, developed by Judkins (OECD, 2002[6]), ensured that the sum of the factors was 3.
Because the nominal sample size for TALIS 2018 was n = 200 schools, a maximum of G =
100 zones or pseudo strata was created for each participating country and a series of G=100
BRR replicate weights were also computed and stored. The creation of BRR weights was
applied to all participating countries/economies, regardless of the size of the sample and of
the method of sample selection.
where 𝜃̂1 and 𝜃̂2 are the two characteristics of interest (e.g. hours paid, hours worked)
measured within each participating school.
The standard error for the difference of the estimates for two countries, say 𝜃̂𝐴 and 𝜃̂𝐷 , is:
1
𝑠𝑒(𝜃̂𝐴 − 𝜃̂𝐷 ) = {𝑉̂𝐹𝐴𝑌 (𝜃̂𝐴 ) + 𝑉̂𝐹𝐴𝑌 (𝜃̂𝐷 )}2 ,
and the standard error for the difference of an estimate for a given country, say 𝜃̂𝐴 and the
international average 𝜃̂̅ , is:
1
̂𝐴 ) + ∑𝐶 𝑉
̂ 𝐹𝐴𝑌 (𝜃
(𝐶 2 − 2𝐶)𝑉 ̂ ̂ 2
̂ ̂
̅ 𝑘=1 𝐹𝐴𝑌 (𝜃𝑘 )
𝑠𝑒 (𝜃𝐴 − 𝜃) = { } ,
𝐶2
where 𝜃̂̅ = ∑ 𝜃̂𝑘 ⁄𝐶 , C is the number of countries contributing to the mean 𝜃̂̅ , and 𝜃̂𝐴 is the
estimate for country A.
Comparisons of subpopulations within a country should be done with a regression on a
dummy variable, as illustrated in the following example. Suppose that the difference
between male and female teachers for some characteristic (e.g. hours of class management
per week) is of interest. We can set a dummy variable Gender = 0 if male, Gender = 1 if
female. A regression model can then be written as Score = a0 + a1 Gender. Clearly, if
Gender = 0, then Scoremale =a0. Likewise, by setting Gender = 1, we obtain Scorefemale =a0
+ a1. Design-based estimation of the regression parameters a0 and a1 can be done using
appropriate software that uses the replicate weights to estimate the standard errors of the
regression parameters. If the test of significance on a1 cannot reject the null hypothesis
H0 : a1 = 0, then we must conclude that scores for male teachers and female teachers are not
significantly different.
If 𝜃̂ is one of the statistics described above and 𝑠𝑒(𝜃̂)is the standard error of 𝜃̂, then we can
easily obtain confidence intervals about zero by computing the following boundaries:
𝑙𝑜𝑤𝑒𝑟𝛼 = 𝜃̂ − 𝑡𝛼,𝑑𝑓 𝑠𝑒(𝜃̂ ) and 𝑢𝑝𝑝𝑒𝑟𝛼 = 𝜃̂ + 𝑡𝛼,𝑑𝑓 𝑠𝑒(𝜃̂) ,
2 2
where 1-α is the pre-set confidence level (e.g. 1-α = 0.95) and 𝑡𝛼,𝑑𝑓 is 1-α/2 percentile of
2
the Student distribution with df degrees of freedom. In most applications, df will be large
enough to allow the use of the standard normal deviate 𝑧1−𝛼 (e.g. 𝑧1−𝛼 = 1.96 for α = 0.05).
2 2
However, in order to confirm the number of degrees of freedom, we still need to verify how
many zones actually contribute to the statistic 𝜃̂ and how many BRR replicates contribute
to the computation of 𝑠𝑒(𝜃̂). This matter is covered in greater detail in the TALIS 2018 and
TALIS Starting Strong Survey 2018 User Guide (OECD, forthcoming[7]).
Alternatively, the design effect can be regarded as the ratio of sample sizes. We can then
speak of “effective sample size” to describe the sample size of the complex survey adjusted
for the design effect:
𝑛
𝑛effective = .
𝑑𝑒𝑓𝑓
The following tables give the estimated design effect for selected key scale variables from
the principal questionnaire (Table 9.2 to Table 9.4) and the teacher questionnaire
(Table 9.5 to Table 9.7), by the participating country/economy. Deff values near 1 mean
that the design was as precise as a simple random sample of the same size (200 schools or
4 000 teachers respectively). Deff values larger than 1 mean that the sampling design was
less efficient than a simple random sample of the same size; deff values smaller than 1
indicate a gain in precision. Alternatively, the deff values indicate by what factor the sample
size was affected: the larger the deff, the smaller the effective sample size; reciprocally, the
smaller the deff, the larger the effective sample size.
Because the value of the deff depends on the design itself (efficiency of the stratification,
clustering, sample size) and on the true sampling variance of the variable in the population,
it varies from one variable to the next.
In Chapter 5 (Table 5.2), a hypothetical deff of 5.2 was used to derive the expected effective
sample size for teachers. The tables presented below show how the actual sample designs
implemented in the various participating countries outperformed the design hypotheses,
thus resulting in effective samples that were much larger than the expected nominal 400
teachers. However, these tables also show that the various samples are not always as
efficient as simple random samples of schools, despite stratification. In all cases,
stratification was implemented to obtain reliable estimates for domains of interest rather
than as a measure to reduce sampling error.
Table 9.2. Estimated design effects for selected scales, ISCED level 2, principal data
1. See endnote 1.
Source: OECD, TALIS 2018 database.
Table 9.3. Estimated design effects for selected scales, ISCED level 1, principal data
Table 9.4. Estimated design effects for selected scales, ISCED level 3, principal data
Participating country/economy Diversity beliefs Organisational innovativeness Overall job satisfaction School leadership
Alberta (Canada) 0.94 1.19 1.65 1.56
Brazil 1.48 0.91 1.12 1.38
Croatia 1.31 1.91 2.36 1.16
Denmark 1.09 1.09 1.06 0.79
Portugal 1.10 0.94 1.29 1.40
Slovenia 1.08 1.29 0.91 1.04
Sweden 2.04 2.04 0.00 3.04
Turkey 2.49 2.84 2.89 2.64
Chinese Taipei 1.29 1.19 1.03 1.73
United Arab Emirates 1.02 0.80 1.15 0.97
Viet Nam 1.65 1.13 1.10 1.29
Table 9.5. Estimated design effects for selected scales, ISCED level 2, teacher data
Participating country/economy Job satisfaction Teacher co-operation Teacher self-efficacy Teaching practices
Alberta (Canada) 1.55 3.08 2.03 4.13
Australia 1.83 1.74 1.48 1.41
Austria 2.56 1.90 1.61 1.33
Belgium 1.67 2.16 1.33 0.88
Flemish Community (Belgium) 1.67 2.16 1.33 0.88
Brazil 2.75 3.66 1.96 1.72
Bulgaria 2.64 2.48 2.06 2.06
Ciudad Autónoma de Buenos Aires 3.27 3.43 2.95 1.86
(Argentina)
Participating country/economy Job satisfaction Teacher co-operation Teacher self-efficacy Teaching practices
Chile 2.00 1.56 1.11 1.87
Colombia 4.01 4.58 3.32 2.26
Croatia 3.28 3.33 2.72 1.63
Cyprus1 2.94 2.47 1.15 1.75
Czech Republic 2.34 2.98 1.50 1.82
Denmark 1.88 2.79 1.67 1.44
England (United Kingdom) 2.09 1.51 1.73 1.61
Estonia 1.79 3.35 1.61 2.15
Finland 2.8.0 2.33 1.81 1.81
France 2.01 2.00 1.11 0.97
Georgia 2.91 3.09 1.72 2.34
Hungary 3.13 2.60 1.83 1.42
Iceland 1.00 1.10 1.10 0.82
Israel 2.34 2.03 1.39 0.82
Italy 2.16 2.31 1.47 1.42
Japan 1.57 2.44 1.38 1.72
Kazakhstan 2.97 3.63 2.50 2.87
Korea 3.05 2.08 1.83 1.36
Latvia 2.89 1.80 1.98 2.12
Lithuania 3.15 2.25 1.85 1.64
Malta 1.79 2.62 1.57 0.73
Mexico 1.88 1.65 1.82 1.27
Netherlands 2.38 8.42 3.97 4.98
New Zealand 3.41 4.06 2.90 1.94
Norway 2.43 3.61 2.29 1.90
Portugal 2.89 2.29 1.28 1.07
Romania 2.57 2.53 2.38 2.12
Russian Federation 2.33 3.19 4.08 3.34
Saudi Arabia 1.75 1.75 1.83 1.31
Shanghai (China) 2.50 2.20 1.32 1.49
Singapore 1.12 0.92 1.05 1.23
Slovak Republic 1.94 1.80 1.35 1.83
Slovenia 1.57 2.17 1.20 1.17
Spain 3.63 4.47 2.82 4.10
Sweden 2.91 3.17 1.59 1.34
Chinese Taipei 2.15 2.25 1.13 1.15
Turkey 3.25 2.33 1.61 1.81
United Arab Emirates 1.39 0.96 0.98 1.52
United States 6.96 6.83 2.44 4.74
Viet Nam 4.37 5.24 3.75 6.75
1. See endnote 1.
Source: OECD, TALIS 2018 database.
Table 9.6. Estimated design effects for selected scales, ISCED level 1, teacher data
Participating country/economy Job satisfaction Teacher co-operation Teacher self-efficacy Teaching practices
Flemish Community (Belgium) 2.00 2.01 1.65 2.23
Ciudad Autónoma de Buenos Aires (Argentina)* 3.17 2.50 2.48 2.19
Denmark 2.04 2.15 1.21 1.71
England (United Kingdom) 2.90 2.38 2.05 1.64
France 2.25 2.80 2.56 2.90
Japan 2.25 2.81 1.77 1.83
Korea 3.18 2.32 1.76 1.65
Spain 3.33 4.67 3.70 3.33
Sweden 2.25 2.10 1.57 1.76
Turkey 5.96 3.62 4.01 2.99
Chinese Taipei 2.07 2.64 1.32 1.35
United Arab Emirates 1.07 1.39 1.10 1.34
Viet Nam 3.76 4.80 4.02 2.42
Table 9.7. Estimated design effects for selected scales, ISCED level 3, teacher data
Participating country/economy Job satisfaction Teacher co-operation Teacher self-efficacy Teaching practices
Alberta (Canada) 2.25 3.42 1.74 1.39
Brazil 2.96 5.67 3.19 1.80
Croatia 1.95 2.47 1.34 2.05
Denmark 2.54 1.66 1.41 1.25
Portugal 2.45 2.60 1.60 1.39
Slovenia 1.17 1.25 1.44 1.22
Sweden 2.48 2.38 1.82 1.80
Turkey 5.78 4.93 5.34 4.13
Chinese Taipei 2.22 2.37 1.60 1.29
United Arab Emirates 1.31 1.11 1.47 1.62
Viet Nam 3.51 4.38 2.64 1.50
References
Fay, R. (1989), “Theory and application of replicate weighting for variance calculation”, [5]
Proceedings of the Section on Survey Research Methods of the American Statistical
Association, pp. 212-217.
Lohr, S. (1999), Sampling: Design and Analysis, Duxbury Press, New York. [1]
OECD (2009), PISA 2006 Technical Report, PISA, OECD Publishing, Paris, [4]
https://dx.doi.org/10.1787/9789264048096-en.
OECD (2002), PISA 2000 Technical Report, OECD Publishing, Paris, [6]
http://dx.doi.org/doi.org/10.1787/19963777.
OECD (forthcoming), TALIS 2018 and TALIS Starting Strong Survey 2018 User Guide, OECD, [7]
Paris.
Rust, K. and J. Rao (1996), “Variance estimation for complex estimators in sample surveys”, [2]
Statistics in Medical Research, Vol. 5/4, pp. 381-397.
Wolter, K. (2007), Introduction to Variance Estimation, 2nd edition, Springer Verlag, New [3]
York.
Notes
1
Note by Turkey: The information in this document with reference to “Cyprus” relates to the
southern part of the Island. There is no single authority representing both Turkish and Greek Cypriot
people on the Island. Turkey recognises the Turkish Republic of Northern Cyprus (TRNC). Until a
lasting and equitable solution is found within the context of the United Nations, Turkey shall
preserve its position concerning the “Cyprus issue”.
Note by all the European Union Member States of the OECD and the European Union: The Republic
of Cyprus is recognised by all members of the United Nations with the exception of Turkey. The
information in this document relates to the area under the effective control of the Government of the
Republic of Cyprus.
2
TIMSS: trends in International Mathematics and Science Study, https://www.iea.nl/timss.
3
See endnote 1.
4
See endnote 1.
This chapter covers the adjudication of TALIS 2018. The first two sections of the chapter
address the purpose of adjudication and what was adjudicated during TALIS 2018, after
which the criteria used to assess each parameter are presented and briefly described in the
third section. The last section describes the recommended usage rating for each
participating country/economy and survey population. Individual rating recommendations
can be found in Annex G.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
The basic principle guiding adjudication was to determine, for each participating
country/economy and for each of the TALIS options, whether the data released to the
countries and to the OECD were “fit for use” as intended.
To establish fitness for use, several quality assurance processes were designed and
activated throughout the survey. Some processes relied on expert advice and opinion; some
relied on qualitative information and learned judgement; and some relied on quantitative
information. The quality observations that the international quality observers (IQOs)
conducted in each participating country/economy provide an example of adjudication in
practice (see Chapter 7).
In general, the consortium considered the overall quality of the survey implementation and
the data yielded to be high. However, several issues arose during the survey with respect
to, for example, national adaptations of the source questionnaires and extensions of the data
collection window. The consortium adjudicators discussed and clarified each issue with the
relevant participating country/economy and sometimes sought advice from the OECD
Secretariat. Once the consortium found a solution that was not only agreeable to the
participating country/economy but also complied with the TALIS 2018 Technical
Standards (see Annex B), it considered the issue settled in terms of having no impact on
data quality. Some problems were discovered only during data processing, weighting or
scaling. The impact of each of these issues on data quality – that is, their potential to limit
the utility of the data – was assessed and documented.
During the adjudication session, held at the IEA Hamburg offices in September 2018 in the
presence of representatives of the OECD Secretariat, each individual dataset, that is, all
data from each country’s/economy’s survey options1 and questionnaire types,2 was
submitted to the same examination. For the first time since the original TALIS survey of
2008, the consortium adjudicated the principal/school data independently of the teacher
data. The members of the consortium responsible for a particular key step of the survey
presented, discussed and assessed any unresolved issues that might reduce each dataset’s
final fitness for use. The following references provide a detailed review of survey processes
and of the principles and implementation of quality assurance in a survey: Kish (1965[2]),
Statistics Canada (2003[3]; 2009[4]; 2017[5]) and the United Nations Commission for Europe
(UNECE, 2014[6]).
Outstanding and unresolved situations likely to diminish the overall utility of a dataset
could occur during any step of the survey process. For example:
Adaptation of questionnaires to the national context: incomplete questionnaires
(national questionnaires needed to include all of the introductory texts and
questions in the international source versions of the questionnaires as well as the
corresponding notes, instructions, response categories and coding schemes);
questions removed or modified without agreement.
Translation and verification: changes to national translations of trend items that
would make comparisons difficult.
Quality of sampling frame (before sampling and confirmation at weighting):
measure of size disproportionate to what was known of the participating
country/economy; missing values.
Handling of out-of-scope and of refusal units: for example, a replaced unit found
to be out of scope.
Handling of replacement and “volunteer” units: for example, replacement schools
participating when the original school was also on the database.
Teacher rosters and within-school sampling: apparent incompleteness; apparent
biased selection of teacher sample.
Main survey administration: noncompliance with the technical standards and
survey operational procedures, units, rules and guidelines; failure to administer the
data collection within the agreed data collection window; failure to administer the
questionnaires according to the tracking form.
Data collection (paper and online modes): missing records (physical or electronic).
Data cleaning: issues with cleaning and editing; need to make post-collection call-
backs to national project managers (NPMs).
Quality observers’ reports: reported issues included differences in the
documentation of national decisions on translation verification feedback and the
actual implementation in the final national instruments; unnatural and non-fluent
national translations; deviations in sampling procedures for listing all eligible
participants in a school; failure to administer the questionnaires according to the
tracking form.
Weighting: miscoded schools or teachers; only rough comparisons of survey results
with frame information; need to make call-backs to NPMs; estimates of population
sizes not matching information on frame; school listings too short; school ID having
no matching unit on frame.
As a general criterion, the consortium members at the adjudication considered that any
problem that had been satisfactorily resolved was no longer a problem and that they did not
need to discuss it. The consortium member mainly responsible for the issue at hand only
presented the issue if unusual circumstances made it stand out.
Otherwise, if a problem had been only partially resolved, the adjudication committee
worked through a series of questions to help them clearly determine the nature of the issue
and its potential impact on data quality: What was the problem? What solutions had been
tried and failed? What solutions had partially worked and to what extent? Was the
perceived impact such that words of caution should be issued to users? Did the adjudication
committee feel that international (or national) comparisons appeared to have been
compromised or limited to the largest subpopulations? Was more information from
countries/economies needed to assess the issue in full?
Once the committee had assessed each survey process, they formulated a recommended
rating for it that accounted for the participation rates and any unresolved issues.
For easy reference, Tables 10.1 and 10.2 reproduce the adjudication tables found in the
TALIS 2018 Technical Standards (see Annex B).
Table 10.1. Adjudication rules for school or principal data in TALIS 2018
School participation
(returned principal questionnaires) (%) Risk of school
Rating
non-response bias
before replacement after replacement
≥ 75 ≥ 75 Good
50–75 ≥ 75 Fair A
50–75 Low Fair C
High Poor D
< 50 Insufficient
≥ 75 ≥ 75 ≥ 75 Good
50–75 Fair A
50–75 ≥ 75 ≥ 75 Fair B
50–75 Low Fair C
High Poor D
50–75 50–75 Poor E
< 50 ≥ 75 Poor F
< 50 < 75 Insufficient
The aim of the following bulleted list is to help data users understand what constitutes
limitations on use or quality of the data:
Good: the participating country’s/economy’s data can be used for all reporting and
analytical purposes and should be included in international comparisons.
Fair (line A): national and subnational estimates can be produced; some teacher
characteristics may be less precise, as indicated by a larger standard error (s.e.),
hence the warning “fair”, but with no additional warnings to users deemed
necessary.
Fair (line B, only for teacher data adjudication): national and subnational estimates
can be produced; some subnational estimates may be of lower precision (larger s.e.)
if the sample size is locally low, hence the warning “fair”, but with no additional
warnings to users considered necessary.
Fair (line C): national and subnational estimates can be produced; some
subnational estimates may be of lower precision (larger s.e.) if the sample size is
locally low, hence the warning “fair”, but with the possible inclusion of a note on
data quality that points to the outcome of the non-response bias analysis (NRBA);
school participation somewhat lower than under (B), meaning that comparison of
subnational estimates needs to be done with care given that some of these results
are based on just a few schools; comparison of small subnational estimates with
similar groups from other countries is unlikely to uncover statistically meaningful
differences because of potentially overly large standard errors.
Poor (line D): in addition to the warnings issued for the previous category, a note
that warns users of indications of non-response bias in some estimates should be
appended; comparisons of subnational estimates need to be limited to the groups
with the larger sample sizes (because the sample at this point represents between
37% and 56% of the teaching workforce, from a relatively small sample of schools,
comparisons with similar groups in foreign countries is inadvisable).
Poor (line E, only for teacher data adjudication): subnational estimates not
recommended; a note pointing out the difficulty of obtaining a representative
sample of schools, therefore, needs to be appended.
Poor (line F, only for teacher data adjudication): limitations similar to those for
line E, but with the inclusion of a note pointing out the difficulty of obtaining at
least 50% participation of the selected sample of schools; evident risk of having a
non-representative sample of schools.
Insufficient: weights should not be calculated for any official tabulations, meaning
that data should not be incorporated3 into international tables, models, averages,
etc.
Thus, the final ratings depended on participation rates before and after replacements, on
data quality issues raised during the adjudication session and on the apparent severity of
the non-response biases. The next six tables present the recommended rating for each
participating country/economy, by ISCED level and population.4
As mentioned earlier, the recommended rating was based on the participation rates
(weighted or unweighted) most favourable to the countries. Detailed results of unweighted
and weighted participation can be found in Annex G.
Principals’
Number of Principals’
Estimated size of participation Recommended
Participating country/economy participating participation after
school population before rating
principals replacement (%)
replacement (%)
Australia* 223 6 522 48.8 77.9 Insufficient
Notes: A school was deemed a participating school if the principal returned his or her questionnaire with at least 1 question
answered.
Australia: data collection window extended into the following school year.
Flemish Community (Belgium): entries on the sampling frame were administrative units and not “schools” as they are usually
defined. Because a “school” may comprise one or several administrative units, the principal would have been reporting for the
school, not just the selected administrative unit. Users, therefore, need to exercise care when analysing the school-level statistics.
Denmark: because non-response bias analysis showed no evidence of a high risk of school non-response bias, Denmark’s rating
was upgraded from “poor” to “fair”. Item PQ-39b and PQ-39c were dropped because of an inaccurate translation.
France: item PQ-06c was withdrawn at France’s request because the wording was not sufficiently clear to ensure
non-misinterpretation of the data. Item PQ-14c was dropped due to an inaccurate translation.
Netherlands: the Netherlands began its data collection six weeks earlier than the other Northern Hemisphere countries/economies
and had an extended collection window. Because the Netherlands Ministry of Education, Culture and Science decided to support
all schools willing to participate, this resulted in the inclusion of some 50 “national” schools that were not included in the
international dataset but were left on the national dataset. Thus, participation rates were computed on the international dataset.
Item PQ-12 was withdrawn at the Netherlands’ request because the public/private status of schools in the Netherlands is not always
obvious and this question was often misinterpreted, despite the explanation provided.
Spain: item PQ-03 was dropped because of an inaccurate translation.
Sweden: item PQ-07a was dropped because of an inaccurate translation.
Source: OECD TALIS 2018 Database.
Notes: A school was deemed a participating school if at least 50% of the selected teachers returned their respective questionnaires
with at least 1 question answered.
Australia: Australia’s data collection window extended into the following school year.
Flemish Community (Belgium): entries on the sampling frame were administrative units and not “schools” as they are usually
defined. Because a “school” may comprise one or several administrative units, the principal would have been reporting for the
school, not just the selected administrative unit. Users, therefore, need to exercise care when analysing the school-level statistics.
France: Item TQ-33b and TQ-33d were dropped because of an inaccurate translation.
Denmark: because non-response bias analysis showed no evidence of a high risk of school non-response bias, Denmark’s rating
was upgraded from “poor” to “fair”.
Korea: in some schools, teacher listings were found to be incorrect; these schools were therefore categorised as “non-participant”.
Netherlands: the Netherlands began its data collection six weeks earlier than the other Northern Hemisphere countries/economies
and had an extended collection window. Because of an unapproved collection protocol that resulted in the inclusion of some 50
“national” schools that were not included in the international dataset but were left on the national dataset, participation rates were
computed on the international dataset.
Spain: item TQ-03 was dropped because of an inaccurate translation.
Source: OECD TALIS 2018 Database.
1. Note by Turkey: The information in this document with reference to “Cyprus” relates to the southern part of the Island. There
is no single authority representing both Turkish and Greek Cypriot people on the Island. Turkey recognises the Turkish Republic
of Northern Cyprus (TRNC). Until a lasting and equitable solution is found within the context of the United Nations, Turkey shall
preserve its position concerning the “Cyprus issue”.
2. Note by all the European Union Member States of the OECD and the European Union: The Republic of Cyprus is recognised
by all members of the United Nations with the exception of Turkey. The information in this document relates to the area under the
effective control of the Government of the Republic of Cyprus. Notes: A school was deemed a participating school if the principal
returned his or her questionnaire.
Alberta (Canada): because non-response bias analysis showed no evidence of a high risk of school non-response bias, Alberta’s
rating was upgraded from “poor” to “fair”. TALIS was conducted as a new Collective Bargaining Agreement for teachers was
being negotiated.
Australia: Australia’s data collection window extended into the following school year.
Austria: item PQ-06 was withdrawn at Austria’s request because the wording was not sufficiently clear to ensure
non-misinterpretation of the data.
Belgium and Flemish Community (Belgium): entries on the sampling frame were administrative units and not “schools” as they
are usually defined. Because a “school” may comprise one or several administrative units, the principal would have been reporting
for the school, not just the selected administrative unit. Users, therefore, need to exercise care when analysing the school-level
statistics.
French Community (Belgium): The sample size is lower than the minimum number of schools required for TALIS, namely, 150,
unless a census of all schools is conducted. Item PQ-14c was dropped due to an inaccurate translation.
Colombia: non-response bias analysis showed no evidence of a high risk of school non-response bias.
Czech Republic: because some translation issues could still exist in the trend items, users need to exercise caution when comparing
results across TALIS cycles.
Denmark: because non-response bias analysis showed no evidence of a high risk of school non-response bias, Denmark’s rating
was upgraded from “poor” to “fair”. Items PQ-39b and PQ-39c were dropped because of an inaccurate translation.
France: item PQ-06c was withdrawn at France’s request because the wording was not sufficiently clear to ensure
non-misinterpretation of the data. Item PQ-14c was dropped due to an inaccurate translation.
Georgia: the overall quality of the translation was found to be questionable. It is also likely that translation issues still exist in the
Georgian and Azerbaijani instruments for Georgia that could affect the data.
Iceland: because Iceland missed 75% participation by only 1 school, Iceland’s rating was upgraded from “poor” to “fair”.
Israel: ultra-orthodox schools were removed post-facto because of very low responses rates, making coverage identical to that of
TALIS 2013. Because translation issues could still exist in the trend items, users need to exercise caution when comparing results
across TALIS cycles. Item PQ-12 was withdrawn at Israel’s request because the classifications of private schools were not defined
well enough to ensure non-misinterpretation of data.
Italy: because translation issues could still exist in the trend items, users need to exercise caution when comparing results across
TALIS cycles. Item PQ-16 was withdrawn at Italy’s request.
Mexico: item PQ-04e, “years working in other jobs”, was dropped because of an inaccurate translation.
Netherlands: the Netherlands began its data collection six weeks earlier than the other Northern Hemisphere countries/economies
and had an extended collection window. Because an unapproved collection protocol resulted in the inclusion of some 50 “national”
schools that were not included in the international dataset but were left on the national dataset, participation rates were computed
on the international dataset. Item PQ-12 was withdrawn at the Netherland’s request because the classifications of private schools
were not defined well enough to ensure non-misinterpretation of data.
New Zealand: coverage was extended to small schools (four or fewer teachers). While the impact of this action on the target
population of teachers was negligible, the impact on the target population of principals is important because, compared to TALIS
2013, the target population for principals nearly doubled in size. TALIS 2018 data comparisons with TALIS 2013 should,
therefore, be restricted to the 2013 coverage.
Russian Federation: Moscow was excluded from TALIS 2018.
Saudi Arabia: two provinces bordering Yemen were excluded from TALIS 2018.
Shanghai (China): item PQ-04d, “years worked as a teacher in total”, was dropped because of an inaccurate translation.
Spain: item PQ-03 was dropped because of an inaccurate translation.
Sweden: item PQ-07a was dropped because of an inaccurate translation.
United Arab Emirates: Comparisons with TALIS 2013 must be limited to Abu Dhabi; data from Abu Dhabi were not adjudicated
independently from those of the UAE. Because of the selection of multi-level schools, the principal data were copied from the
original ISCED level 2 principal questionnaire to the ISCED level 1 and ISCED level 3 corresponding forms, except for item PQ-
17.
Source: OECD TALIS 2018 Database.
Italy: because translation issues could still exist in the trend items, users need to exercise caution when comparing results across
TALIS cycles.
Netherlands: the Netherlands began its data collection six weeks earlier than the other Northern Hemisphere countries/economies
and had an extended data collection window. Because an unapproved collection protocol resulted in the inclusion of some 50
“national” schools that were not included in the international dataset but were left on the national dataset, participation rates were
computed on the international dataset.
New Zealand: coverage was extended to small schools (four or fewer teachers). While the impact on the target population of
teachers was negligible, the impact on the target population of principals is important because, compared to TALIS 2013, the
target population of teachers nearly doubled in size. Comparison of TALIS 2018 data with TALIS 2013 data should, therefore, be
restricted to the 2013 coverage.
Norway: item TQ-42p was withdrawn on Norway’s request because of a problematic national adaptation that could have led to
misinterpretation of the data.
Russian Federation: Moscow was excluded from TALIS 2018. Item TQ-10 b was dropped because of an inaccurate translation.
Item TQ-34 was dropped because of an error in the layout of the questionnaires that could have led to misinterpretation of the
data.
Saudi Arabia: two provinces bordering Yemen were excluded.
Shanghai (China): item TQ-17 was dropped because of an inaccurate translation.
Spain: item TQ-03 was dropped because of an inaccurate translation.
United Arab Emirates: Comparisons with TALIS 2013 must be limited to Abu Dhabi; data from Abu Dhabi were not adjudicated
independently from those of the UAE. Because of the selection of multi-level schools, the principal data were copied from the
original ISCED level 2 principal questionnaire to the ISCED level 1 and ISCED level 3 corresponding forms, except for item PQ-
17.
Source: OECD TALIS 2018 Database.
Notes: A school was deemed a participating school if the principal returned his or her questionnaire.
Alberta (Canada): because non-response bias analysis showed no evidence of a high risk of school non-response bias, Alberta’s
rating was upgraded from “poor” to “fair”. TALIS was conducted as a new Collective Bargaining Agreement for teachers was
being negotiated.
Denmark: because non-response bias analysis showed no evidence of a high risk of school non-response bias, Denmark’s rating
was upgraded from “poor” to “fair”. Items PQ-39b and PQ-39c were dropped because of an inaccurate translation.
Portugal: Part B of item PQ-15 was dropped because of an inaccurate translation.
Slovenia: Slovenia missed 75% participation by 0.5% of a principal, which led to the recommendation to upgrade Slovenia’s rating
from “poor” to “fair”.
Sweden: item PQ-07a was dropped because of an inaccurate translation.
United Arab Emirates: Comparisons with TALIS 2013 must be limited to Abu Dhabi; data from Abu Dhabi were not adjudicated
independently from those of the UAE. Because of the selection of multi-level schools, the principal data were copied from the
original ISCED level 2 principal questionnaire to the ISCED level 1 and ISCED level 3 corresponding forms, except for item PQ-
17.
Notes: A school was deemed a participating school if at least 50% of the selected teachers returned their respective questionnaires.
Alberta (Canada): because non-response bias analysis showed no evidence of a high risk of school non-response bias, Alberta’s
rating was upgraded from “poor” to “fair”. TALIS was conducted as a new Collective Bargaining Agreement for teachers was
being negotiated.
United Arab Emirates: Comparisons with TALIS 2013 must be limited to Abu Dhabi; data from Abu Dhabi were not adjudicated
independently from those of the UAE. Because of the selection of multi-level schools, the principal data were copied from the
original ISCED level 2 principal questionnaire to the ISCED level 1 and ISCED level 3 corresponding forms, except for item PQ-
17.
Source: OECD TALIS 2018 Database.
Sampling of schools for the TALIS-PISA link is described in Chapter 5 of this report. In
summary, a sample of 150 schools (unless discussions with the NPM led to a different size)
was drawn randomly from the sample of schools drawn for PISA. As the PISA data
collection proceeded, the set of (original sample or replacement) schools participating in
PISA emerged, and thus the set of schools where the TALIS-PISA link should be
administered. The distribution of the TALIS Principal Questionnaire and of the TALIS
Teacher Questionnaire to a sample of twenty “PISA teachers” (i.e. teachers of 15-year-old
students) could then go ahead in each of the schools that had participated in PISA and also
sampled for the TALIS-PISA link.
The school and teacher participation rates for the TALIS-PISA link are computed as they
are for the ISCED levels (see Chapter 9 for details). Since the TALIS-PISA link sample is
a random subsample of the PISA sample of schools, the TALIS-PISA school weights (or
school weight component of the teacher weight) refer back to the original PISA population.
The adjudication of the TALIS-PISA samples had to wait until the PISA samples had been
adjudicated, as the former was dependent on the latter to allow the final determination of
the recommended rating. Even if the recommended rating for TALIS-PISA link, solely
based on what happened during the preparation and collection of the TALIS-PISA link,
were “good”, if the data or samples from PISA ware to be rated less favourably, the
matched file could not be adjudicated as “good”. It could only be adjudicated as the
weakest, at most, of either rating.
Table 10.9 and Table 10.10 display the participation rates for the principals and teachers in
each country/economy that participated in the TALIS-PISA Link.
Principals’
Number of Number of Number of Principals’
Participating participation Recommended
schools sub- eligible participating participation after
country/economy before rating
sampled schools principals replacement (%)
replacement (%)
Australia 150 148 131 66.9 88.5 Fair
Ciudad Autónoma de
Buenos Aires 104 81 77 88.9 95.1 Good
(Argentina)
Colombia 162 162 153 91.4 94.4 Good
Czech Republic 190 182 175 96.8 96.8 Good
Denmark 150 150 83 52.0 57.8 Poor
Georgia 150 144 124 86.1 86.1 Good
Malta 63 50 47 94.0 94.0 Good
Turkey 150 147 142 96.6 96.6 Good
Viet Nam 150 115 115 100.0 100.0 Good
Australia 150 148 131 2 233 34 598 65.6 88.8 93.4 82.9 Good
Ciudad
Autónoma de
104 81 73 1 194 2 673 87.7 90.1 85.1 76.7 Good
Buenos Aires
(Argentina)
Colombia 162 162 154 2 242 179 900 91.4 95.1 94.6 89.9 Good
Czech
190 182 173 2 592 62 040 95.1 95.1 94.8 90.0 Good
Republic
Denmark 150 150 100 1 079 20 777 65.8 70.0 85.9 60.2 Poor
Georgia 150 144 132 1 923 24 592 93.1 93.1 94.3 87.8 Good
Turkey 150 147 142 3 591 236 904 97.9 97.9 99.6 97.5 Good
Viet Nam 150 115 114 2 170 250 645 99.3 99.3 98.4 97.7 Good
References
OECD (2018), Teaching and Learning International Survey (TALIS) 2018 (database), OECD, Paris. [8]
OECD (2017), TALIS 2018 Technical Standards, Internal Document, OECD, Paris. [7]
OECD (2015), Call for Tenders 100001191: Teaching and Learning International Survey (TALIS 2018), [1]
OECD, Paris,
http://www.oecd.org/callsfortenders/2015%2002%2025%20CFT%20for%20TALIS%202018.pdf.
Statistics Canada (2017), Statistics Canada’s Quality Assurance Framework, 3rd Edition, Statistics [5]
Canada, Ottawa, https://www150.statcan.gc.ca/n1/en/pub/12-586-x/12-586-x2017001-
eng.pdf?st=YIabaxa0.
Statistics Canada (2009), Quality Guidelines, 5th Edition, Statistics Canada, Ottawa, [4]
https://unstats.un.org/unsd/econstatkb/knowledgebasearticle10247.aspx.
Statistics Canada (2003), Survey Methods and Practice, Statistics Canada, Ottawa, [3]
https://www150.statcan.gc.ca/n1/en/catalogue/12-587-X.
UNECE (2014), Generic Statistical Business Process Model, United Nations Economic Commission for [6]
Europe,
https://www.unece.org/fileadmin/DAM/stats/documents/ece/ces/ge.57/2015/Paper_CES_GSPBM_EN.
pdf.
Notes
1
Survey option refers to ISCED level 1, ISCED level 2, and ISCED level 3. The TALIS-PISA link
data could not be adjudicated at that time.
2
Questionnaire type refers to the teacher questionnaire or the principal questionnaire.
3
At their last meeting held in Paris in November 2018, the technical advisory group recommended
that data from countries that had not reached 50% participation be nonetheless weighted and
displayed in tables but not used in the computation of international averages or models.
4
Table 10.3 to Table 10.8 display the participation-rate estimates that were the most favourable for
the adjudication rating. The most favourable estimates could have been weighted or unweighted
depending on the characteristics of the country/economy, the teacher and principal population and
the educational level.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
11.1. Overview
The TALIS questionnaires include numerous items pertaining to, for example, school
characteristics and principals’ and teachers’ attitudes, perceptions and practices. Use of
suitable statistical procedures allows for the combination of responses to these items into
indices or scales. In line with previous TALIS cycles, two types of combinations were
considered for TALIS 2018:
Simple indices (e.g. ratios), constructed through simple arithmetical
transformations or by recoding one or more items.
Scale scores, derived using latent modelling within the framework of confirmatory
factor analysis (CFA).
CFA treats items as indicators of unobservable personal characteristics of respondents, such
as self-efficacy or beliefs, and uses combinations of items to develop a model of the latent
construct. After testing and confirming a stable latent construct model, scale scores may be
estimated that serve as numeric values for the latent constructs.
While simple indices enhance the analysis of observable (manifest) characteristics, such as
student-teacher ratio, scales enable analysis of non-observable (latent) characteristics such
as attitudes or other personal traits. This chapter begins by outlining the procedures used to
compute simple indices. It then describes the procedures involved in scale evaluation and
scale score estimation. A detailed description of the scales, including the items used to
compose each scale and the results of scale evaluation, follows. The chapter ends by
exploring the implications of these results for further analyses, especially in relation to
cross-country/economy comparisons.
This section describes the construction of simple indices that are used in multiple tables of
the OECD final report (OECD, 2019[1]) and are part of the publicly available international
database. These indices were constructed through arithmetical transformation or recoding
of one or more items. More details are provided in Chapter 12.
Student-teacher ratio
The student-teacher ratio was calculated at the school-level based on the information
derived from school principals’ responses to questions about the number of currently
employed teachers1 (headcounts) and the total number of enrolled students (headcounts)
from all grades. Thus, the index reflects the overall student-teacher ratio in each school
rather than being restricted to the target population. The ratio (STRATIO) is derived by
dividing the total number of students enrolled (TC3G16) by the number of employed
teachers in a given school (TC3G13A).
and personnel for pedagogical support in each school rather than being restricted to the
target population. The ratio (TPRATIO) was derived by dividing the number of teachers
(TC3G13A) by the number of personnel for pedagogical support (TC3G13B).
School autonomy: School autonomy for staffing (T3PAUTS); School autonomy for
budgeting (T3PAUTB); School autonomy for educational policies (T3PAUTP),
School autonomy for instructional policies (T3PAUTI); School autonomy for
curriculum (T3PAUTC)
To describe the extent of school autonomy in decision making, indices were derived from
question TC3G20 of the school principal questionnaire. Five simple indices were formed:
school autonomy for staffing (T3PAUTS), school autonomy for budgeting (T3PAUTB),
school autonomy for educational policies (T3PAUTP), school autonomy for instructional
policies (T3PAUTI), and school autonomy for curriculum (T3PAUTC).
School autonomy indices were created using 11 items, each of which had five response
options. School principals had to indicate who, among a range of stakeholders, had a
considerable responsibility in making decisions relating to tasks listed in the questionnaire.
Considerable responsibility could be attributed to one or more of the following: the
principal, other members of the <school management team>, teachers (not a part of the
<school management team>), <school governing board> or <local, municipality/regional,
state, or national/federal>. For a particular task, the extent of school-level autonomy was
determined by whether considerable responsibility lay at the school level (i.e. with the
principal, other members of the <school management team>, teachers (not as a part of the
<school management team>) and the <school governing board>), with other authorities
(i.e. <local, municipality/regional, state, or national/federal>) or was shared by both groups.
Each response option (checked/not checked) was a variable of its own. The 11 items
describing the tasks produced 55 variables in total. Table 11.1 lists the indices with the
corresponding items.
checked, then the new variable was coded as -1. Thus, if the principal checked only
the “other authority” response option, the task was considered to be an external
responsibility (not autonomous).
3. If the school principal checked response options from both groups (decision making
at the school level and decision making by other authorities), the responsibility was
considered to be shared, and the value remained 0 (see point 1).
4. If one of the first four response options (the decision making being at the school
level) was checked, and the fifth response option (decision making being other
authority responsibility) was not checked, the new variable was coded as +1.
Therefore, if the principal selected at least one of the four school-level
responsibility options and no other authority responsibility option, the task was
considered to be a school responsibility (autonomous).
5. The newly created variables were recoded: -1 was recoded to 1, 0 to 2, +1 to 3.
6. For each index, if more than half of the newly created variables were classified as
autonomous, the school was classified as autonomous. If more than half of the
corresponding tasks were classified as not autonomous, the school was classified
as not autonomous. If neither criterion was met, the school was classified as mixed.
The final indices were coded 1 for “no autonomy”, 2 for “mixed autonomy”, and 3
for “autonomy”.
TALIS aims to collect robust and rich information about teachers’ and principals’
characteristics as well as about their schools. Many of the specific personal traits, for
instance, teachers’ attitudes and self-efficacy, cannot be observed directly, but only through
expressed opinions or intended and observed behaviour. Such traits are considered latent,
that is, not directly observable. In large-scale studies, sets of items drawn from the studies’
instruments are used to estimate these latent traits. The instruments used in the TALIS
surveys are the teacher and principal questionnaires, the items of which are designed to
reflect specific facets of the envisaged latent traits. The procedure used to combine
responses to these questions into a single scale score representing the latent construct of
interest is called scaling.5
analysis and interpretation), the omega coefficients from the scale’s subscales were used
to calculate the stratified coefficient alpha as follows:
∑𝑖 𝜎𝑖2 (1 − 𝛾𝑖 )
𝛾𝑆𝑇𝑅𝐴𝑇,𝛼 = 1 −
𝜎𝑐2
where 𝛾𝑆𝑇𝑅𝐴𝑇,𝛼 is the reliability of the composite/multidimensional scale; 𝜎𝑖2 is the variance
of the ith subscale, or the variance of the standardised factor scores of this subscale; 𝛾𝑖 is
the reliability of the ith subscale or the omega for this subscale; and 𝜎𝑐2 is the variance of
the composite/multidimensional scale or the variance of the sum of the standardised factor
scores of the subscales (He, 2010[9]).
Model analysis
Confirmatory factor analysis (CFA) was conducted to evaluate how well the actual
empirical data reflected the predefined latent construct. CFA allows inference on the scale
from the scale items by establishing associations between the two. The association between
each item and the scale is described in a regression line (OECD, 2014[10]). Using the model
fit indices as evaluation criteria (for further details, see Table 11.3), the specified (or
theoretical) model of each scale is assessed with respect to its alignment to the empirical
data (Hu and Bentler, 1999[11]).
Various procedures can be used to estimate the scale scores, including computation of a
sum or mean score over all items that measure the same construct, computations based on
classical test theory (CTT), structural equation modelling (SEM) using CFA, and
computing person parameters based on item response theory (IRT). The results from these
methods are typically highly correlated but are not completely congruent, and each method
has its respective advantages and disadvantages.
In keeping with past TALIS cycles, scale score computation based on CFA was used during
the current cycle of TALIS (OECD, 2014[10]; OECD, 2010[12]), as it remains a method that
has a solid scientific basis and offers great flexibility, given the invariance results for
TALIS scales. Also, when employed with certain modelling software (Mplus8), this form
of computation is well equipped to deal with missing values.
Analysis during the TALIS 2018 cycle was based on the general SEM framework, where
CFA is a specific type of model classified within this framework (Schreiber et al., 2006[13]).
All constructs with ordinal response categories were scaled using continuous CFA
(estimated using robust full-information maximum likelihood estimator on the matrix of
Pearson’s correlations), while constructs9 with binary items were scaled using categorical
CFA modelling (estimated using robust WLSMV10 estimator on the pair-wise matrix of
tetrachoric correlations). Design weights and replicate weights were used for all analyses,
and weights were rescaled so that each country contributed equally to the estimates.11 Some
TALIS 2018 items had already been used in previous TALIS cycles to construct latent scale
scores. Item selection for TALIS 2018 was conducted to maximise the overlap of items
between TALIS 2018 and TALIS 2013 and thus allow for item level comparisons across
cycles.
Given the increased emphasis on measurement invariance testing during the TALIS 2018
cycle (see below) compared to the previous cycles of TALIS, model fit for the current cycle
was prioritised over comparability with the earlier cycles. Therefore, directly comparing
the scale scores from TALIS 2018 with those of the past cycles is not recommended.
and that the content validity of the scale was therefore low. Decisions made at this stage
were discussed with the TALIS Consortium and the QEG.
During the second step, the CFA model was tested via use of country/economy-level data
in each of the analysed populations (in all countries/economies for ISCED levels 1, 2 and
3, as applicable14). Improvements implemented in the pooled models were applied to these
models. In total, 72 single-country/economy, single-ISCED level models were analysed for
each scale.15 The evaluation procedure for these models was very similar to that of the
pooled model. If a model at the single-country/economy, single-ISCED level failed to meet
the fit index cut-off criteria (Table 11.3), additional improvements for that specific
population were implemented. If one of these scale models could not be improved, then
that single-country/economy, single-ISCED level population was removed from further
parameter estimation and included for the scale score construction that used fixed
parameters from the final scale model.
Table 11.3 provides information on the cut-off criteria for the CFA model fit evaluation
(Brown, 2015[15]; Chen, 2007[16]; DeVellis, 2003, pp. 94-96[17]; Hoyle, 2014[18]). It is
important to stress that these statistical criteria were used for the decision-making process
that was based on an iterative process involving content-related considerations between the
IEA scaling team and the QEG members.
Table 11.3. Cut-offs for CFA model evaluation for TALIS 2018
Notes: Cronbach’s alpha, omega, and stratified coefficient alpha are all different estimators of the same
reliability/internal consistency; therefore, the criteria are the same.
The cut-off criterion for the SRMR was less strict in TALIS 2013 (SRMR≤.1). To enhance alignment with the
established cut-off criteria for model evaluation proposed in the literature (Hu and Bentler, 1999[11]; OECD,
2014[10]; Schermelleh-Engel, Moosbrugger and Müller, 2003[19]; Steiger, 1990[20]; Yu, 2002[21]), a stricter cut-
off of SRMR was applied during TALIS 2018. This cut-off was justified because the scale construction in
TALIS 2018 benefitted from prior TALIS cycles as well as from the field trial.
When the intercepts are fixed to either a very high or low value, the SRMR can be misleading. In these cases,
the variances may be very low, resulting in an extremely large SRMR value (Asparouhov and Muthén, 2018[22]).
Thus, decisions relating to the performance and the measurement invariance of the scale are primarily based on
CFI, TLI, and RMSEA. SRMR was used for model evaluation if other fit indices revealed inconsistent results
and in models with varying intercepts (configural and metric models).
groups, which assumes comparability across groups. The degree to which the stricter model
(i.e. a model that assumes equal parameters across groups), as compared to the less strict
model (i.e. a model that assumes some degree of flexibility of parameters between groups)
suits the data is evaluated via model fit indices and the direction (i.e. better or worse fit)
and degree of change between the fit indices of each model.
In the least restrictive model, all parameters are freely estimated for each group17 (e.g.
participating country/economy) separately, meaning that the parameters are unconstrained
and vary across these groups. The model implies that there is no comparability between
groups because all parameters are group-specific and therefore no statistical comparisons
are permitted. Additional models are estimated and become gradually more restrictive (i.e.
contain a greater number of parameters that are restricted to be equal across groups),
resulting in increasingly equal measurement models with greater levels of comparability
between the groups. As the models become more restrictive, the justification for statistical
comparability and analysis between the groups (e.g. correlation or mean comparisons)
increases. The levels of comparability are called measurement invariance levels and are
specifically defined by the parameters that are restricted in the model representing each of
the levels. The three most common levels of measurement invariance testing are (1)
configural, (2) metric and (3) scalar (Cheung, 1999[23]; Cheung and Rensvold, 2002[4];
Davidov, 2008[24]; Davidov et al., 2014[27]; Steenkamp and Baumgartner, 1998[25]; OECD,
2014[10]).
The lowest level of measurement invariance, the configural level, assumes that the
underlying latent construct is specified by a particular configuration of items in all analysed
groups in the same way. Configural level of measurement invariance applies when the
construct is measured by the same items. It implies that the structure of the construct
indicated by the configuration of items is equivalent across participating
countries/economies. If a scale reaches only the configural level of measurement
invariance, then any statistical method applied to compare the scale scores across groups
will violate the basic assumption of the comparability of the measured construct. The
comparability occurs at a conceptual level only, while score comparability is not achieved.
Therefore, results (e.g. correlations) from different groups can be discussed only through
reference to each specific group.
The meaning of the scale is defined by the content of the questions participants were asked
and that were used to create the scale. If the strength of the associations (i.e. the magnitude
of the regression parameters) is the same across groups, then the latent construct is assumed
to have the same meaning. This is the second level of measurement invariance, the metric
level. Metric level of measurement invariance applies when (1) the structure of the
construct is the same across groups, and (2) the strength of the associations between the
construct and the items (factor loadings) is equivalent across groups. Metric invariance
makes it possible to claim that one unit of change in the construct will lead to the same
amount of average change in the items that constitute the construct across different groups
(e.g. participating countries/economies). If a scale establishes the metric level of
invariance, it can be assumed that comparisons of correlational analyses (such as
correlation or regression analysis) are free of the cross-group bias. Of note, for scales with
binary indicators this level of measurement invariance testing was omitted as these models
have identification issues when using the Mplus software.
During TALIS 2018, the level of comparability was deemed sufficient if all parameters of
a model (except the residuals of the items18) were fixed to be the same across groups (e.g.
participating countries/economies). This approach is called the scalar level of measurement
invariance. Scalar level of measurement invariance applies when (1) the structure of the
construct is the same across groups, (2) the strength of the associations between the
construct and the items (factor loadings) are equivalent, and (3) the intercepts/thresholds
for all items across groups are equivalent. If the intercepts of the items for all groups are
equivalent, then the expected value of the items becomes the same across groups when the
value of the construct is zero, meaning that the value/degree of the construct for a certain
value of the observed item can be claimed to be equivalent across different groups. In this
case, cross-group comparisons of scale means are justified, and the results can be assumed
to be free of cross-group bias (e.g. cross-cultural bias).
The current cycle of TALIS sought to use up-to-date and valid techniques with solid
analytical backgrounds for the scaling procedure to ensure the resulting model of each scale
was an accurate representation of teachers’ and principals’ characteristics in the
participating countries/economies. This aim resulted in an examination of the measurement
invariance across both participating countries/economies and ISCED levels.
The measurement invariance testing referred to as “cross-country/economy” examined
invariance within a single ISCED level. In other words, for the participating
countries/economies that participated at the ISCED 1 level, measurement invariance testing
was conducted considering each participating country/economy with an ISCED level 1
population was considered to be a separate group during the analysis. This same procedure
was followed for each ISCED level (ISCED 1, 2 and 3) separately. The measurement
invariance testing referred to as “cross-ISCED level” examined invariance within a single
country/economy across ISCED levels. Therefore, measurement invariance testing was
conducted for each participating country/economy that participated in the study at more
than one ISCED level, which meant that the invariance testing treated each ISCED level a
separate group during the analysis. The measurement invariance testing both cross-country
and cross-ISCED level was then used to develop each scale’s final model containing
parameter constraints as suggested by the results of both the cross-country and cross-
ISCED level measurement invariance testing conducted for that particular scale.
The modelling method chosen to investigate measurement invariance during TALIS 2018
was the same as that used during for TALIS 2013 and TALIS 2008, namely the multiple
group confirmatory factor analysis (MGCFA). During the cross-country measurement
invariance testing, the analysis evaluated whether the model was invariant/equal across
participating countries/economies within a certain ISCED level.19 These models were
compared at configural, metric, and scalar levels, and the purpose of the analyses was to
investigate if statistical analysis of the scale scores could be compared across participating
countries/economies within each ISCED level.
During the cross-ISCED level measurement invariance testing, the analysis evaluated
whether the model was invariant/equal across ISCED levels within a participating
country/economy. To be specific, for a certain country X, up to three20 CFA models were
created and compared at configural, metric, and scalar levels. The purpose of this analysis
was to investigate whether statistics obtained from the analysis of the scale scores could be
compared across ISCED levels within a single country/economy. The scale score
estimation was based on the evaluation of each scale’s results as described in the following
sections.
The changes to the model fit indices CFI, TLI, RMSEA and SRMR/WRMR21 were used to
evaluate the measurement invariance level of each scale, and the criteria used to conduct
the evaluation were as follows:
For those participating countries/economies that did not meet the TALIS 2018 technical
standards,14 those participating countries/economies with late data submission, and the
TALIS-PISA link countries/economies, their respective final scale models used fixed
parameters. The parameters were fixed according to the cross-country measurement
invariance results of ISCED level 2, the TALIS study’s target population. Therefore, if a
scale reached scalar invariance cross-country within the ISCED 2 level, then the factor
loadings and intercepts were constrained to be equal to those unstandardised parameters in
the final scale model; if metric invariance was reached, only factor loadings were
constrained; and if configural invariance was reached, then no constraints were imposed on
factor loadings and intercepts.17
Once the final scale models had been specified for all participating countries/economies
and ISCED levels, factor score determinacies from the model were used to calculate the
omega reliability coefficient as part of the evaluation of the scale. All parameters estimated
in these models are reported below in the section Results from scales evaluation and scale
score creation. For scales reaching metric or scalar levels of invariance within a certain
ISCED level, unstandardised factor loadings and intercepts are presented to show the
model’s equality restraints. In addition, standardised factor loadings are shown to aid in
interpretation of the reliability of the model.22
According to the SEM framework, an item 𝑦 is predicted from the latent factor 𝜂, which is
multiplied with the matrix of factor loadings 𝜦. The vector of item intercepts 𝝉 and the
vector of residuals 𝜀 are both added to the product. This is written as:
𝑦 = 𝝉𝑦 + 𝚲 𝑦 𝜂 + 𝜀
To estimate factor models from ordinal items, the MLR estimation procedure for
continuous latent constructs was used because it is robust to non-normality. Mplus uses the
maximum of the posterior distribution of the factor, which is known as the maximum a
posteriori method (Muthén and Muthén, 1998-2017[30]). This method is similar to the latent
regression approach (Skrondal and Laake, 2001[31]). If all 𝑦 items are continuous, the factor
score estimate 𝜂 for individual 𝑖 is based on a regression method with correlated factors
(Muthén, 1977[32]), where the factor score is computed from the mean vector of 𝑦 items,
denoted as 𝝁, the factor score coefficient matrix 𝑪, the vector of observations 𝒗𝑖 , the vector
of intercepts 𝝉, and the matrix of factor loadings 𝜦 multiplied by the mean vector 𝝁:
𝜂̂ 𝑖 = 𝝁𝑦 + 𝑪(𝒗𝑖 − 𝝉𝑦 − 𝚲𝑦 𝝁𝑦 )
The factor score coefficient matrix, in turn, is based on the item covariance matrix 𝜮, the
matrix of factor loadings 𝜦, and the matrix of residual variances and covariances 𝜣:
𝐶 = 𝚺𝑦 𝚲𝑇𝑦 (𝚲𝑦 𝚺𝑦 𝚲𝑇𝑦 + 𝚯𝑦 )−1
These formulas imply that higher factor loadings on an item are associated with a stronger
influence of this item on the factor score estimate. Likewise, the larger the residual variance
of an item, the smaller its influence on the factor score estimate. The factor loadings, item
intercepts, the mean vector and the variance of the latent variable affect the estimated
scores.
The WLSMV estimation procedure was used to estimate factor models with scaled binary
items. This method produces weighted least square parameter estimates by using a diagonal
weight matrix, robust standard errors, and a mean- and variance-adjusted χ2 test statistic
(Brown, 2006, p. 388[14]). The method also takes a slightly different approach to estimating
factor scores. First, the probability of observed binary response 1 is defined as
1
−
𝑓𝑗 (𝑦𝑖𝑗 |𝜼𝑖 ) = Φ [(𝜁 − 𝝀𝑗′ 𝜼𝑖 )𝜃𝑗𝑗 2 ]
and the probability of observed categorical response 0 is therefore 1 − 𝑓𝑗 (𝑦𝑖𝑗 |𝜂𝑖 ), where 𝜁
is the item threshold based of an item 𝑗, 𝝀𝑗′ is the 𝑗th row of the matrix of factor loadings
𝜦, and 𝜃𝑗 is the 𝑗th diagonal of the matrix of residual variances and covariances 𝜣, while 𝜼𝑖
is a vector of true factor scores.
The factor score estimates 𝜂̂ 𝑖 are then found as the mode of the posterior distribution of 𝜼𝑖
by minimising, through use of quasi-Newton techniques, the following function 𝐹 with
respect to 𝜼𝑖 :
𝑝
1
𝐹 = (𝜼𝑖 − 𝝁𝑖 )′ 𝚺 −1 (𝜼𝑖 − 𝝁𝑖 ) − ∑ ln 𝑓𝑗 (𝑦𝑖𝑗 |𝜼𝑖 )
2
𝑗=1
where 𝝁 is the mean vector of 𝑦 items. Contrary to the factor score estimation for models
with categorical items, this approach assumes uncorrelated residual variances even if
residual covariances are allowed.
For both continuous and categorical data, Mplus provides a model-based approach to
estimating parameters in a model with missing data. Model-based approaches account for
the missing data and estimate the missing parameters in one step (Lüdtke et al., 2007[33]),
and to do this Mplus uses the expectation maximisation algorithm. For a detailed
description see Dempster, Laird and Rubin (1977[34]). This procedure assumes that the data
are missing-at-random, meaning the probability of a missing observation depends not on
the true score of a person regarding the latent construct but can be correlated with other
covariates of the scale (Schafer and Graham, 2002[35]).
For each TALIS scale, the expectation maximisation algorithm was used to compute a scale
score for respondents who responded to at least one of the items belonging to the respective
scale. (The algorithm made it possible to deal with missing data and provide the appropriate
estimator for the continuous or categorical nature of the respective scales.) The residual
variances of the items were allowed to be freely estimated in all models.
For the standardisation procedure of the scale score, the estimated scale scores were
standardised using data from the target population of the TALIS study, ISCED level 2 (with
the exception of participating countries/economies not meeting the technical standards and
participating countries/economies with late data collection; see Table 11.7). A metric with
a standard deviation of 2.0 and a mean of 10 was used to standardise the scale scores. The
mathematical transformation ensured that all or almost all values were positive, thus
allowing for a convenient interpretation.
Once the scores were adjusted with a standard deviation of 2.0 and mean of 10, a second
adjustment was made as follows. The average scale score for the set of those individuals
from ISCED level 2 whose IMV was equal to the midpoint of the scale items (in this
example, 2.5) was computed and then subtracted from 10, and this difference (𝐹̅𝑀∗ ) was
added to the scale score of each individual, resulting in the final standardised scores.
Mathematically, the standardisation is represented as follows:
𝐹𝑖 − 𝐹̅ ∗
𝑋𝑖 = 10 + 2 + 𝐹̅𝑀∗
𝜎𝐹∗
where 𝑋𝑖 is the standardised scale score of individual 𝑖, 𝐹𝑖 is the raw estimated scale score
of that individual, 𝐹̅ ∗ is the mean scale score of the ISCED level 2 level population, 𝜎𝐹∗ is
the standard deviation of the scale score of the ISCED level 2 population, and 𝐹̅𝑀∗ is as
described above.
By adding 𝐹̅𝑀∗ , standardised scale scores are shifted so that the scale score average is 10 for
those individuals from the target population with an IMV equal to the midpoint of the
response scale. This shifts the scale score mean to 10 plus the constant 𝐹̅𝑀∗ and allows for
easy interpretation of scale score: scores above 10 suggest positive associations with the
scale (e.g. agreement, of more importance), scores below 10 suggest negative associations
with the scale (e.g. disagreement, of lesser importance), while scores of exactly 10 suggest
indifference.
Table 11.4 provides an example of this procedure for the scale T3PERUT. The column on
the left shows all the possible IMVs for individuals when aggregating the responses to the
four items of the scale. The column on the right shows the mean factor score of individuals
from the target population with the corresponding IMV. As shown, individuals with an
IMV of 2.5 have an average scale score of 10. In general, these means rise as the IMV rises,
and fall as the IMV falls.23
Table 11.4. Items average scale score equivalent table for the scale T3PERUT
subscales. The computation of the composite scores for individual 𝑖 can be summarised as
follows:
∑𝑁
𝑗=1 𝑋𝑖𝑗
𝑌𝑖 = ,
𝑁
where 𝑌𝑖 are the composite scores for a certain multidimensional scale for individual 𝑖, 𝑋𝑖𝑗
are the raw scale scores of the subscale 𝑗 for individual 𝑖, and 𝑁 is the number of subscales
of this multidimensional scale. These scores were then standardised in the same manner as
described in sub-section Scale score standardisation.
Any analysis of the subscale scores and the composite scores needs to take into account the
following limitations: (1) the subscale scores should not be used in a correlation or
regression analysis simultaneously due to collinearity of these subscales; and (2) the
composite scores could be biased because of the weight of the subscale on the latent
construct not being taken into account (i.e. assumed to be equal for all subscales).
Cross-country/economy comparability
An important consideration for anyone using the scale scores in analysis is that the results
not only from the scale reliability analyses but also the cross-country or ISCED-level
measurement invariance testing described above have major implications for (1) the
construction of scale scores, and (2) the use of the scale scores in further analysis.
To aid the user regarding the comparability of scale scores, the cross-country levels of
invariance are included in the variable labels of each scale in the international data sets. For
example, the scale with variable name T3CLASM has the label, “Classroom management
/ Metric (1) - Configural (2, 3)”, which indicates that the scale reached metric invariance
for ISCED 1 level and configural invariance for ISCED level 2 and 3. Table 11.5 shows
how many scales reached a particular level of invariance. For a more detailed look, Table
11.6 presents the specific invariance levels for each scale, listed by its variable name and
label found in the international database.
Table 11.5. Scale counts of the invariance levels for both populations
Invariance levels
Population Configural Metric Scalar
Teachers
ISCED 1 6 23 2
ISCED 2 8 22 1
ISCED 3 9 20 2
Principals
ISCED 1 5 6 1
ISCED 2 6 5 1
ISCED 3 5 6 1
Table 11.6. Invariance level reached for each scale by ISCED level
In addition, recommendations for analysis based on the different levels of invariance are
provided below. The recommendations are specific to cross-country invariance (i.e. within
a single ISCED level) but may also be applied to cross-ISCED level invariance within a
single participating country/economy. Different levels of measurement invariance provide
different potentials for the analysis of data and reporting. The proposed analyses for each
level of measurement invariance are:
Cross-country analysis of scales with only configural level of invariance:
o Recommendation for analysis: At the cross-national level, only qualitative
(descriptive) comparisons are statistically justified (e.g. “associations of
staff-beliefs and staff education is positive in country A and B, whereas
there is no significant association in country C”), which should be
presented together with the limitations concerning the interpretation of the
results, in particular concerning the differences between participating
countries/ economies with respect to the meaning of the construct.
o Limitation: If a scale only reaches configural invariance, the scale score is
constructed in such a way that the factor loadings and intercepts are
allowed to vary across participating countries/economies. Further analysis
aimed at cross-country comparisons can only be conducted at the
conceptual level meaning that no statistical methods of comparison (such
as t-tests) are applicable in such cases.
Cross-country analysis of scales with metric level of invariance:
o Recommendation for analysis: For such a scale, the strength of the
association between the scale and items is comparable across participating
countries/economies, and statistical methods such as correlation and
regression are applicable. Comparisons of associations between
participating countries/economies are justified (e.g. “The association of
staff-beliefs and level of education are significantly stronger in
countries/economies A and B than in country/economy C”). The metric
level of measurement invariance means scales can be used for analysis
based on correlation and linear regression.
o Limitation: If a scale reaches only the metric level of invariance the score
of the scale is created respectively with equal factor loadings but with
intercepts allowed to vary across participating countries/economies, and
therefore a mean score between participating countries/economies is not
possible.
Cross-country analysis of scales with scalar level of invariance:
o Recommendation for analysis: For these scales, the comparison of the
mean score of the scale is meaningful across groups. Scale mean can be
compared across participating countries/economies (e.g. “Staff-beliefs are
significantly higher in country A compared to countries/economies B
and C”).
Violating these assumptions or ignoring these recommendations may or will lead to biased
results and/or interpretations. For instance, if scale scores based on the assumption of
metric invariance are created and used to compare the country/economy means of that
scale, the differences between participating countries/economies will most likely be biased.
Thus, it is recommended that data users apply statistical procedures to analyse scale scores
accordingly to its invariance level reached.
participating countries/economies contributed equally to the analysis – see the TALIS 2008
Technical Report 2008 (OECD, 2010, p. 139[12]).
composite multidimensional scales. Table 11.3 presents the criteria guiding interpretation
of each scale’s reliability.
The information on reliability is followed by a description of the scale-specific results from
the model estimation and comparisons. Confirmatory factor analysis (CFA) was used to
evaluate the theoretically developed scale models reflecting the study’s latent constructs,
while the model fit indices presented at this point were those obtained from the model
analysis conducted at the country/economy level for each scale (i.e. the CFA conducted
separately for each of the analysed populations at the country/economy level).27 The criteria
used for model evaluation and selection appear in the first part of the chapter (Table 11.3).
Not reported are the model fit indices for the scales based on three items.28
The results of the measurement invariance testing across participating countries/economies
within each ISCED level follows the section on reliabilities (a summary of the results from
the measurement invariance testing across ISCED levels within each participating
country/economy appear in the Annex I tables.) The results of the model comparisons of
measurement invariance testing were used to establish the invariance levels of each scale.
Bold font is used to denote the invariance level of each scale in the tables.
The next part of the scale description centres on the international unstandardised model
parameters that were estimated via the scales’ final models. The factor loadings and
intercepts (the parameters that were uniform across participating countries/economies) are
reported for the scales that reached scalar invariance. However, only the factor loadings are
reported for the metric invariant scales. For those scales that reached only configural
invariance unstandardised item parameters are country specific and, therefore, not reported.
In addition, the country-level standardised factor loadings are reported for each
participating country per ISCED level in a separate table because they were standardised
at the country level, meaning that these standardised parameters differed at the country
level even for the scalar invariant scales. The factor loadings indicate the strength of the
relationship between each item and the latent scale. The values were interpreted in
accordance with the cut-off criteria provided in Table 11.3. The standardised factor
loadings are presented next to the unstandardised intercepts. The intercepts indicate the
predicted values for an item when the value of the latent trait was zero.
Four multidimensional scales were included in the teacher population and one in the
principal population. The multidimensional scales were evaluated with the same model fit
criteria used for the unidimensional scales (Table 11.3). The fit indices indicated acceptable
model fits in all cases. The composite scales29 were computed by averaging the scores from
the subscales.30 Therefore, in addition to the overall model evaluation of the
multidimensional constructs, their respective subscales were evaluated separately. The
measurement invariance level for an averaged scale index was determined by the lowest
invariance level of its respective subscales.
The international parameters from the final scale models were used to estimate the scale
scores for: (1) those participating countries/economies that did not meet the requirements
determined in the TALIS technical standards (Australia and the Netherlands for
ISCED level 1 for all scales and, in addition, Australia, ISCED level 2, for principals
scales only;31 (2) those participating countries/economies that collected their data later than
the designated time; and (3) the TALIS-PISA link populations32 (Australia, Ciudad
Autónoma de Buenos Aires [Argentina], Colombia, the Czech Republic, Denmark,
Georgia, Malta, Turkey and Viet Nam). These parameters included fixed factor loadings in
the ISCED level 2 metric models, and fixed factor loadings and intercepts in the ISCED
level 2 scalar models. At times, additional participating countries/economies were excluded
from the scale evaluation because of model non-convergence, instability or some other
issue (Table 11.7). Because these additional countries/economies were excluded during the
model analyses at the country/economy level from further evaluation, they did not
contribute to the final scale models and there are no scale scores for these scales. This case
only pertains to scales from the principal questionnaire. Reporting on these excluded
populations is provided, where applicable, in a separate section for excluded populations.
Table 11.7. Excluded populations from the estimation of the parameters of principal scales
Table 11.8. Item wording for teacher motivation and perceptions scales
Table 11.9. Reliability coefficients for teacher motivation and perceptions scales
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. Note by Turkey
The information in this document with reference to “Cyprus” relates to the southern part of the Island. There is
no single authority representing both Turkish and Greek Cypriot people on the Island. Turkey recognises the
Turkish Republic of Northern Cyprus (TRNC). Until a lasting and equitable solution is found within the context
of the United Nations, Turkey shall preserve its position concerning the “Cyprus issue”.
Note by all the European Union Member States of the OECD and the European Union
The Republic of Cyprus is recognised by all members of the United Nations with the exception of Turkey. The
information in this document relates to the area under the effective control of the Government of the Republic
of Cyprus.
Source: OECD, TALIS 2018 database.
the cut-off value for all populations except Bulgaria, Georgia, Hungary, Japan, Korea,
Lithuania, Malta, the Russian Federation and the Slovak Republic for ISCED level 2, Japan
and the Netherlands for ISCED level 1, and Denmark, Malta and Viet Nam for the TALIS-
PISA link populations. Finally, the SRMR is above the cut-off criterion in South Africa for
ISCED level 2, Australia and the Netherlands for ISCED level 1, and for all the TALIS-
PISA link populations except those in Australia and the Czech Republic.
The scales T3SOCUT and T3VALP were both constructed from only three items, which
means the model fit indices for them are not reported here (see endnote 28). The only
exceptions where the model fit statistics of models with three items could be estimated
pertained to those populations for which model parameters were fixed, specifically, those
countries that did not meet the TALIS technical standards. These countries were Australia
and the Netherlands for ISCED level 1, South Africa for ISCED level 2 because of the
delay in its data submission, and all TALIS-PISA link populations. Because the models in
these populations were specified by fixing the estimation parameters to be equal to the
international parameters, it was possible to estimate model fit.33 As evident in Table 11.11
and Table 11.12, the model fit statistic for these populations is inconsistent: in some
populations the scale models perform well (e.g. the Netherlands’ ISCED level 1 population
for the scale T3SOCUT), whereas others do not meet some of the fit statistic cut-off criteria
(e.g. Australia’s ISCED level 1 population for the scale T3SOCUT).
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
Source: OECD, TALIS 2018 database.
Table 11.16. Unstandardised factor loadings for T3PERUT, T3SOCUT and T3VALP
for all countries for all populations
The item parameters in the next three tables are reported for each scale per participating
country/economy. The standardised factor loadings indicate the strength of the relationship
between each item and the scale construct (see the section Description of scales and their
parameters). As presented in Table 11.17, the standardised factor loadings for the scale
T3PERUT are above 0.450 for all populations, which indicate at least a moderate
relationship between the items and the latent construct. More specifically, most values for
items TT3G07A, TT3G07B and TT3G07C are above 0.600, and the lowest item loadings
found for item TT3G07D are still above 0.450 for all populations.
The standardised factor loadings for the scale T3SOCUT presented in Table 11.18 were all
above 0.450. In most populations and for most items, the loadings were above 0.700,
indicating a strong relationship between items and the latent construct.
All factor loadings for all items in the scale T3VALP (Table 11.19 were greater than 0.450.
The only exception was item TT3G54D in Brazil, ISCED levels 1 and 3.
Table 11.17. Standardised factor loadings and unstandardised intercepts for scale T3PERUT
New Zealand 0.737 0.874 0.867 0.518 2.954 3.080 3.129 2.865
Norway 0.734 0.887 0.898 0.563 2.479 2.741 2.907 2.829
Portugal 0.856 0.888 0.897 0.622 2.903 2.811 2.839 2.738
Romania 0.696 0.786 0.794 0.563 3.252 2.984 2.969 3.163
Russian Federation 0.611 0.757 0.776 0.549 2.591 2.968 3.312 3.176
Saudi Arabia 0.693 0.812 0.774 0.496 3.475 3.582 3.574 3.324
Shanghai (China) 0.791 0.812 0.864 0.597 3.520 3.369 3.454 3.374
Singapore 0.768 0.906 0.886 0.548 3.132 3.291 3.255 3.036
Slovak Republic 0.713 0.824 0.821 0.581 2.076 2.506 2.919 2.917
Slovenia 0.875 0.919 0.904 0.607 2.520 2.492 2.577 2.268
South Africa2 0.751 0.762 0.777 0.525 3.320 3.013 3.216 3.069
Spain 0.856 0.933 0.885 0.612 2.831 2.758 2.676 2.705
Sweden 0.702 0.816 0.809 0.524 2.443 2.743 2.779 2.722
Chinese Taipei 0.837 0.890 0.854 0.600 3.576 3.599 3.542 3.503
Turkey 0.572 0.784 0.752 0.549 2.802 3.202 3.282 3.277
United Arab Emirates 0.756 0.770 0.807 0.530 3.477 3.240 3.401 3.309
United States 0.779 0.872 0.881 0.557 3.174 3.110 3.169 3.129
Viet Nam 0.698 0.681 0.657 0.527 3.576 3.138 3.175 3.414
ISCED level 1
Australia1 0.810 0.911 0.895 0.565 3.102 3.088 3.157 2.683
Flemish Community (Belgium) 0.867 0.909 0.823 0.590 2.888 2.993 2.751 2.716
Ciudad Autónoma de Buenos 0.817 0.897 0.886 0.628 2.317 2.199 2.065 2.136
Aires (Argentina)
Denmark 0.785 0.877 0.875 0.566 2.059 2.337 2.303 2.716
England (United Kingdom) 0.772 0.911 0.906 0.508 3.116 3.208 3.270 2.772
France 0.863 0.906 0.868 0.583 2.881 2.822 2.744 2.658
Japan 0.563 0.893 0.894 0.515 2.452 3.130 3.185 2.677
Korea 0.666 0.853 0.900 0.628 2.975 3.132 3.359 3.287
Netherlands1 0.873 0.912 0.859 0.615 2.048 2.069 1.984 1.890
Spain 0.870 0.936 0.889 0.651 2.831 2.758 2.676 2.705
Sweden 0.709 0.838 0.813 0.538 2.443 2.743 2.779 2.722
Chinese Taipei 0.820 0.878 0.841 0.603 3.576 3.599 3.542 3.503
Turkey 0.572 0.793 0.786 0.593 2.802 3.202 3.282 3.277
United Arab Emirates 0.769 0.756 0.814 0.537 3.477 3.240 3.401 3.309
Viet Nam 0.700 0.673 0.653 0.543 3.576 3.138 3.175 3.414
ISCED level 3
Alberta (Canada) 0.814 0.895 0.820 0.472 3.323 3.331 3.318 2.919
Brazil 0.803 0.861 0.863 0.549 3.055 2.843 2.989 2.834
Croatia 0.613 0.883 0.874 0.560 2.627 2.626 2.707 2.562
Denmark 0.753 0.849 0.859 0.539 2.059 2.337 2.303 2.716
Portugal 0.855 0.890 0.901 0.611 2.903 2.811 2.839 2.738
Slovenia 0.845 0.894 0.908 0.591 2.525 2.541 2.556 2.470
Sweden 0.729 0.840 0.815 0.514 2.443 2.743 2.779 2.722
Chinese Taipei 0.841 0.885 0.834 0.603 3.576 3.599 3.542 3.503
Turkey 0.587 0.775 0.740 0.556 2.802 3.202 3.282 3.277
United Arab Emirates 0.773 0.786 0.819 0.539 3.477 3.240 3.401 3.309
Viet Nam 0.730 0.701 0.683 0.545 3.576 3.138 3.175 3.414
TALIS-PISA link
Australia 0.802 0.918 0.914 0.543 3.125 3.167 3.198 2.891
Ciudad Autónoma de Buenos 0.821 0.899 0.855 0.587 2.381 2.265 2.107 2.228
Aires (Argentina)
Colombia 0.827 0.904 0.872 0.568 3.102 2.961 2.968 2.825
Czech Republic 0.756 0.835 0.828 0.545 2.594 2.665 2.652 2.933
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.18. Standardised factor loadings and unstandardised intercepts for scale T3SOCUT
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.19. Standardised factor loadings and unstandardised intercepts for scale T3VALP
Participating countries/economies Standardised factor loadings Unstandardised intercepts
TT3G54C TT3G54D TT3G54C TT3G54D TT3G54C TT3G54D
ISCED level 2
Alberta (Canada) 0.872 0.833 0.710 2.258 2.309 2.349
Australia 0.875 0.830 0.732 2.114 2.194 2.136
Austria 0.722 0.737 0.717 1.928 1.635 1.622
Belgium 0.842 0.750 0.696 1.755 1.818 1.803
Flemish Community (Belgium) 0.807 0.690 0.640 1.970 2.059 2.112
Brazil 0.750 0.437 0.613 1.514 2.552 1.704
Bulgaria 0.832 0.787 0.722 1.884 1.880 1.817
Ciudad Autónoma de Buenos Aires (Argentina) 0.833 0.599 0.708 1.576 1.962 1.596
Chile 0.754 0.461 0.736 1.576 2.324 1.527
Colombia 0.775 0.580 0.694 1.828 2.471 1.951
Croatia 0.863 0.710 0.754 1.395 1.663 1.536
Cyprus 0.843 0.750 0.712 2.087 2.190 1.984
Czech Republic 0.793 0.737 0.642 1.748 1.762 1.934
Denmark 0.819 0.745 0.695 1.454 1.676 1.661
England (United Kingdom) 0.876 0.855 0.715 1.701 1.701 1.796
Estonia 0.770 0.744 0.668 2.026 2.120 2.055
Finland 0.829 0.827 0.676 2.043 2.056 2.416
France 0.788 0.753 0.706 1.589 1.596 1.542
Georgia 0.799 0.766 0.713 2.304 2.256 2.227
Hungary 0.864 0.873 0.750 1.732 1.699 1.642
Iceland 0.773 0.543 0.662 1.543 2.023 1.554
Israel 0.797 0.551 0.692 1.851 2.396 1.911
Italy 0.745 0.547 0.661 1.555 2.059 1.690
Japan 0.880 0.882 0.739 1.788 1.736 1.667
Kazakhstan 0.789 0.766 0.697 2.220 2.385 2.604
Korea 0.847 0.854 0.752 1.978 1.863 1.815
Latvia 0.803 0.822 0.652 1.802 1.844 2.094
Lithuania 0.795 0.715 0.556 1.518 1.671 1.883
Malta 0.826 0.665 0.715 1.763 1.985 1.704
Mexico 0.796 0.517 0.703 1.580 2.392 1.714
Netherlands 0.702 0.630 0.571 1.934 2.106 2.203
New Zealand 0.847 0.713 0.674 1.834 2.092 1.964
Norway 0.834 0.802 0.680 2.053 2.046 2.017
Portugal 0.766 0.473 0.672 1.451 2.120 1.575
Romania 0.772 0.687 0.721 1.989 2.166 2.045
Russian Federation 0.826 0.851 0.759 2.052 2.023 2.058
Saudi Arabia 0.783 0.654 0.714 2.060 2.500 2.234
Shanghai (China) 0.895 0.818 0.830 2.492 2.387 2.524
Singapore 0.857 0.835 0.727 2.424 2.322 2.518
Slovak Republic 0.786 0.702 0.607 1.521 1.734 1.858
Slovenia 0.793 0.591 0.655 1.506 1.786 1.606
South Africa2 0.713 0.677 0.724 2.119 2.444 2.017
Spain 0.796 0.592 0.630 1.512 1.757 1.785
Sweden 0.822 0.822 0.630 1.750 1.715 1.830
Chinese Taipei 0.812 0.870 0.709 1.885 1.780 1.944
Turkey 0.803 0.591 0.737 1.750 2.311 1.721
United Arab Emirates 0.854 0.831 0.770 2.705 2.635 2.826
United States 0.844 0.769 0.722 1.907 2.212 2.102
Viet Nam 0.678 0.656 0.643 2.852 2.998 3.077
“Thinking about your teaching in the <target class>, how often do you do the
following?” (TT3G42), which was followed by items on instruction that were used
for the subscale Clarity of instruction (T3CLAIN), items on student tasks for the
subscale Cognitive activation (T3COGAC), and items on engaging students in the
lesson for the subscale Classroom management (T3CLASM), as presented in Table
11.20.
These three subscales formed the multidimensional scale Teaching practices, composite
(T3TPRA).
population in Viet Nam for T3CLASM. However, low omega values are evident in many
participating countries for the scale T3COGAC, especially with respect to the ISCED level
2 population and for the ISCED level 1 and ISCED level 3 populations. The coefficients
for the composite scale T3TPRA indicate a high level of reliability in all populations.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. Calculation of the omega coefficient was based on the unidimensional models for every single subscale of the
multidimensional construct.
Source: OECD, TALIS 2018 database.
Clarity of instruction
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Cognitive activation
for all ISCED levels. Although the subscale T3CLASM reaches configural invariance for
ISCED levels 2 and 3 only, it meets the criteria for metric invariance for ISCED level 1.
Based on the lowest invariance level of its subscales, the composite scale T3TPRA can be
considered metric invariant for ISCED level 1 and configural invariant for ISCED levels 2
and 3.
Tables 11.29, 11.30 and 11.31 present the standardised factor loadings and unstandardised
intercepts for the subscales T3CLAIN, T3COGAC and T3CLASM respectively. For
T3CLAIN, the factor loadings for items TT3G42B, TT3G42C and TT3G42D are generally
high and all above 0.450. However, item TT3G42BA shows lower factor loadings in many
participating countries (also seen in Table 11.28 above), suggesting that this item is only
weakly related to the scale construct.
In general, the factor loadings for items included in the subscale T3COGAC are above
0.450 in most populations. The strongest relationship observed is that between item
TT3G42F and the latent construct, whereas most of the factor loadings for item TT3G42G
are between 0.450 and 0.600.
In most populations, the factor loadings for T3CLASM for all items are higher than 0.600,
indicating a strong relationship between the items and the latent construct. However,
several populations exhibit lower loadings for items TT3G42K and TT3G42L, as observed
in Latvia for ISCED level 2, and for item TT3G42L in Portugal for ISCED levels 2 and 3.
Table 11.29. Standardised factor loadings and unstandardised intercepts for scale T3CLAIN
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
items TT3G33D, TT3G33E and TT3G33F of subscale T3EXCH are mostly above 0.600,
but this is not the case for item TT3G33G, where most of the factor loadings are below
0.600 but still above 0.450. Although most factor loadings for subscale T3COLES are only
moderate, they are still above the cut-off criterion for all items.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. Omega coefficient was calculated based on unidimensional models for every single subscale of the
multidimensional construct.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. The poor fit of the model affected the TLI calculation, which is not reported.
Source: OECD, TALIS 2018 database.
Note: Although the change in TLI slightly exceeded the cut-off criterion, a decision was made to accept the
metric level of measurement invariance because the fit indices for the metric model were acceptable and the
change in RMSEA met the cut-off criterion.
Source: OECD, TALIS 2018 database.
Table 11.38. Unstandardised factor loadings for T3COLES for all countries for all
populations
T3COLES (Metric)
TT3G33A 0.934
TT3G33B 0.763
TT3G33C 0.853
TT3G33H 0.658
Table 11.39. Standardised factor loadings and unstandardised intercepts for scale T3EXCH
ISCED level 2 populations. Model fit was also poor for the Flemish Community (Belgium)
ISCED level 1 population and the Czech Republic and Georgia TALIS-PISA link
populations.
The model fit results for the scale T3PDIV presented in Table 11.45 suggest a perfect model
fit for most populations, as this scale was measured by just three items. However, model
fit was poor for the Netherlands ISCED level 1 population and the Columbia, Turkey and
Viet Nam TALIS-PISA link populations.
The results for the scale T3PDBAR presented in Table 11.46 exhibit acceptable fit for many
populations. However, compared to other Feedback and development scales, the model fit
for this scale was poor for the following populations: Bulgaria, Columbia, Cyprus,
Denmark, Malta, Norway and Saudi Arabia ISCED level 2; Denmark ISCED level 1; Brazil
ISCED level 3; and Columbia TALIS-PISA link.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection
Source: OECD, TALIS 2018 database.
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. As the correction factor for this country/economy was negative, only the SRMR is reported.
Source: OECD, TALIS 2018 database.
Table 11.51. Unstandardised factor loadings for T3PDPED and T3PDIV for all countries for
all populations
Table 11.52. Standardised factor loadings and unstandardised intercepts for scale T3EFFPD
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.53. Standardised factor loadings and unstandardised intercepts for scale T3PDPED
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.54. Standardised factor loadings and unstandardised intercepts for scale T3PDIV
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.55. Standardised factor loadings and unstandardised intercepts for scale T3PDBAR
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
from the table, the subscales T3SECLS, T3SEINS and T3SEENG have high reliabilities
for most populations. The reliability coefficients are slightly lower (below 0.700),
however, in the Netherlands ISCED level 2 population for the scale T3SEINS, while some
of the omega values for scale T3SEENG are below 0.700 for several populations: Austria,
the Czech Republic, the Netherlands and Norway for ISCED level 2; and Denmark and
Sweden for ISCED level 3.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Self-efficacy in instruction
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.64. Unstandardised factor loadings for T3SECLS, T3SEINS and T3SEENG
for all countries for all populations
Table 11.65. Standardised factor loadings and unstandardised intercepts for scale T3SECLS
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.66. Standardised factor loadings and unstandardised intercepts for scale T3SEINS
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.67. Standardised factor loadings and unstandardised intercepts for scale T3SEENG
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. Omega coefficient was calculated based on unidimensional models for every single subscale of the
multidimensional construct.
4. These participating countries’/economies’ reliability coefficient estimation failed in the final scale model
due to a negative residual variance for one or more items that could not be corrected; when this occurs for a
subscale of a multidimensional scale, the multidimensional scale reliability coefficient is also missing; these
countries/economies have untrustworthy scale scores for the corresponding ISCED level.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.76. Unstandardised factor loadings for T3JSENV, T3JSPRO and T3SATAT
for all countries for all populations
Table 11.77. Standardised factor loadings and unstandardised intercepts for scale T3JSENV
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.78. Standardised factor loadings and unstandardised intercepts for scale T3JSPRO
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.79. Standardised factor loadings and unstandardised intercepts for scale T3SATAT
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
The results for the scale T3WLOAD are presented in Table 11.89. Here, the factor loadings
for item TT3G52B are above 0.600 for all populations, and the same can be said for items
TT3G52A and TT3G52C for most populations. While the factor loadings for item
TT3G52D are above 0.450 for most populations, the factor loading is lower for a good
number of populations. The factor loadings for item TT3G52E are weak for almost all
populations, suggesting a weak relationship between this item and the latent construct.
The factor loadings for the scale T3STBEH item TT3G52G are above 0.600 in all
populations, as depicted in Table 11.80. However, the factor loadings for item TT3G52H
are moderate (between 0.450 and 0.600) in many populations and below 0.450 in some
populations, while the factor loadings for item TT3G52F are below 0.450 for nearly half
the populations. These results suggest the scale is a weak construct in a fair number of
populations.
Table 11.80. Item wording for workplace well-being and stress scales
Table 11.81. Omega coefficients for workplace well-being and stress scales
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. These participating countries’/economies’ reliability coefficient estimation failed in the final scale model
due to a negative residual variance for one or more items that could not be corrected; these countries/economies
have untrustworthy scale scores for the corresponding ISCED level.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Workload stress
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.87. Unstandardised factor loadings for T3WELS, T3WLOAD and T3STBEH for all
countries for all populations
Table 11.88. Standardised factor loadings and unstandardised intercepts for scale T3WELS
1. Data from participating country/economy rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.90. Standardised factor loadings and unstandardised intercepts for scale T3STBEH
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
“How strongly do you agree or disagree with the following statements about this
<target class>?” (TT3G41). The stem was followed by items regarding student
behaviour that were used to form the scale Teacher’s perceived disciplinary climate
(T3DISC).
“How strongly do you agree or disagree with the following statements about what
happens in this school?” (TT3G49). Items about teacher-student interaction
following this question were used to form the scale Teacher-student relation
(T3STUD).
“How strongly do you agree or disagree with these statements, as applied to this
school?” (TT3G48), followed by items concerning school decision making that
were used to form the scale Participation among stakeholders (T3STAKE).
These scales are presented in Table 11.91.
between 0.450 and 0.600 in a fair number of populations, while the loading for the
Denmark ISCED level 2 population is below 0.450.
Lastly, for scale T3STAKE, presented in Table 11.102, most factor loadings for all items
are above 0.600, with notable exceptions for several items for a number of ISCED level 1
populations and for item TT3G48E for some populations.
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Teacher-student relations
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.99. Unstandardised factor loadings for T3DISC, T3STUD and T3STAKE
for all countries for all populations
Table 11.100. Standardised factor loadings and unstandardised intercepts for scale T3DISC
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.101. Standardised factor loadings and unstandardised intercepts for scale T3STUD
1. Data from the participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.104. Omega coefficients for the populations in each participating country/economy
T3TEAM
Participating countries/economies
Omega coefficient
ISCED level 2
Alberta (Canada) 0.912
Australia 0.895
Austria 0.887
Belgium 0.882
Flemish Community (Belgium) 0.876
Brazil 0.933
Bulgaria 0.918
Ciudad Autónoma de Buenos Aires (Argentina) 0.920
Chile 0.925
Colombia 0.935
Croatia 0.931
Cyprus 0.897
Czech Republic 0.882
Denmark 0.887
England (United Kingdom) 0.880
Estonia 0.880
Finland 0.882
France 0.878
Georgia 0.903
T3TEAM
Participating countries/economies
Omega coefficient
Hungary 0.922
Iceland 0.903
Israel 0.903
Italy 0.916
Japan 0.901
Kazakhstan 0.812
Korea 0.920
Latvia 0.924
Lithuania 0.910
Malta 0.865
Mexico 0.920
Netherlands 0.821
New Zealand 0.893
Norway 0.794
Portugal 0.901
Romania 0.939
Russian Federation 0.880
Saudi Arabia 0.908
Shanghai (China) 0.953
Singapore 0.904
Slovak Republic 0.906
Slovenia 0.912
South Africa2 0.876
Spain 0.916
Sweden 0.891
Chinese Taipei 0.916
Turkey 0.956
United Arab Emirates 0.920
United States 0.889
Viet Nam 0.920
ISCED level 1
Australia1 0.906
Flemish Community (Belgium) 0.891
Ciudad Autónoma de Buenos Aires (Argentina) 0.941
Denmark 0.882
England (United Kingdom) 0.922
France 0.918
Japan 0.906
Korea 0.925
Netherlands1 0.899
Spain 0.931
Sweden 0.908
Chinese Taipei 0.908
Turkey 0.958
United Arab Emirates 0.922
Viet Nam 0.941
ISCED level 3
Alberta (Canada) 0.906
Brazil 0.937
Croatia 0.929
Denmark 0.874
T3TEAM
Participating countries/economies
Omega coefficient
Portugal 0.910
Slovenia 0.899
Sweden 0.901
Chinese Taipei 0.918
Turkey 0.951
United Arab Emirates 0.925
Viet Nam 0.912
TALIS-PISA link
Australia 0.903
Ciudad Autónoma de Buenos Aires (Argentina) 0.914
Colombia 0.910
Czech Republic 0.904
Denmark 0.891
Georgia 0.927
Malta 0.889
Turkey 0.929
Viet Nam 0.916
Team innovativeness
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.107. Unstandardised factor loadings and intercepts for T3TEAM for all countries
for all populations
T3TEAM (Scalar)
Unstandardised factor loadings Unstandardised intercepts
TT3G32A 0.574 2.990
TT3G32B 0.592 2.900
TT3G32C 0.598 2.951
TT3G32D 0.559 2.979
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
“In this school, are the following practices in relation to diversity implemented?”
(TT3G47), followed by items regarding schools and diversity that were used to
form the scale Teacher diversity practices (T3DIVP).
Table 11.109 provides information on each scale.
T3SEFE T3DIVP
Participating countries/economies
Omega coefficient Cronbach’s alpha
ISCED level 2
Alberta (Canada) 0.823 0.650
Australia 0.812 0.666
Austria 0.748 0.607
Belgium 0.736 0.669
Flemish Community (Belgium) 0.790 0.684
Brazil 0.773 0.703
Bulgaria 0.746 0.713
Ciudad Autónoma de Buenos Aires (Argentina) 0.696 0.679
Chile 0.792 0.691
Colombia 0.760 0.699
Croatia 0.801 0.676
Cyprus 0.805 0.716
Czech Republic 0.810 0.597
Denmark 0.723 0.660
England (United Kingdom) 0.785 0.663
Estonia 0.771 0.667
Finland 0.716 0.620
France 0.650 0.586
Georgia 0.821 0.712
Hungary 0.733 0.649
Iceland 0.839 0.624
Israel 0.759 0.685
Italy 0.563 0.636
Japan 0.808 0.665
Kazakhstan 0.824 0.624
Korea 0.880 0.737
T3SEFE T3DIVP
Participating countries/economies
Omega coefficient Cronbach’s alpha
Latvia 0.753 0.606
Lithuania 0.834 0.690
Malta 0.771 0.672
Mexico 0.743 0.688
Netherlands 0.799 0.628
New Zealand 0.787 0.449
Norway 0.787 0.594
Portugal 0.667 0.691
Romania 0.774 0.600
Russian Federation 0.815 0.690
Saudi Arabia 0.848 0.751
Shanghai (China) 0.869 0.714
Singapore 0.861 0.635
Slovak Republic 0.733 0.640
Slovenia 0.661 0.563
South Africa2 0.721 0.698
Spain 0.719 0.677
Sweden 0.741 0.585
Chinese Taipei 0.841 0.691
Turkey 0.766 0.790
United Arab Emirates 0.806 0.738
United States 0.805 0.702
Viet Nam 0.699 0.566
ISCED level 1
Australia1 0.801 0.639
Flemish Community (Belgium) 0.781 0.624
Ciudad Autónoma de Buenos Aires (Argentina) 0.753 0.662
Denmark 0.766 0.632
England (United Kingdom) 0.812 0.619
France 0.623 0.476
Japan 0.812 0.665
Korea 0.876 0.606
Netherlands1 0.828 0.609
Spain 0.750 0.687
Sweden 0.740 0.616
Chinese Taipei 0.841 0.665
Turkey 0.753 0.815
United Arab Emirates 0.774 0.732
Viet Nam 0.789 0.529
ISCED level 3
Alberta (Canada) 0.796 0.625
Brazil 0.787 0.710
Croatia 0.780 0.690
Denmark 0.746 0.630
Portugal 0.686 0.715
Slovenia 0.691 0.608
Sweden 0.736 0.628
Chinese Taipei 0.845 0.718
Turkey 0.773 0.812
United Arab Emirates 0.812 0.769
Viet Nam 0.634 0.602
T3SEFE T3DIVP
Participating countries/economies
Omega coefficient Cronbach’s alpha
TALIS-PISA link
Australia 0.801 0.637
Ciudad Autónoma de Buenos Aires (Argentina) 0.736 0.641
Colombia 0.830 0.704
Czech Republic 0.790 0.652
Denmark 0.778 0.628
Georgia 0.783 0.752
Malta 0.757 0.711
Turkey 0.780 0.787
Viet Nam 0.691 0.555
1. Data from participating country/economy rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from participating country/economy rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Diversity practices
1. Data from participating country/economy rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.115. Unstandardised factor loadings for T3SEFE for all countries for ISCED level 2
T3SEFE (Metric)
TT3G45A 0.482
TT3G45B 0.519
TT3G45C 0.583
TT3G45D 0.544
TT3G45E 0.522
Table 11.116. Standardised factor loadings and unstandardised intercepts for T3SEFE
1. Data from participating country/economy rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.117. Standardised factor loadings and unstandardised intercepts for T3DIVP
1. Data from participating country/economy were rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Teaching practices
How often do you use the following methods of assessing student learning in the <target class>?
Response options: “Never or almost never” (1), “Occasionally” (2), “Frequently” (3), “Always” (4)
TT3G43A I administer my own assessment
TT3G43B I provide written feedback on student work in addition to a <mark, i.e. numeric score or letter grade>
TT3G43C I let students evaluate their own progress
TT3G43D I observe students when working on particular tasks and provide immediate feedback
(1931[36])). As a consequence, there is not a reliability coefficient for this population for
this scale in Table 11.120, nor are there standardised parameters in Table 11.121.
The Portugal ISCED level 2 population was excluded from the final scale modelling for
the scale T3PWLOAD because specific parameters within each of these populations caused
model instability. Sweden’s ISCED level 2 population was excluded from the model after
model analysis at the country/economy level due to a non-converge issue. These
populations, therefore, do not have reliability coefficients for this scale in XX11.2 or
standardised parameters in XX11.11. In addition, XX11.5 does not have fit statistics for
Sweden ISCED level 2.
invariant at the ISCED level 3 level. Therefore, the invariance level for the composite scale
T3PJOBSA was configural for all ISCED levels.
Finally, Table 11.125 presents the invariance results for the scale T3PWLOAD. It reached
configural invariance for ISCED level 2 and metric invariance for ISCED levels 1 and 3.
Because the configural models were perfect, if the metric model exhibited acceptable fit,
the model was considered metric invariant.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. Omega coefficient was calculated based on unidimensional models for every single subscale of the
multidimensional construct.
4. These participating countries’/economies’ reliability coefficient estimation failed in the final scale model
due to a negative residual variance for one or more items that could not be corrected; when this occurs for a
subscale of a multidimensional scale, the multidimensional scale reliability coefficient is also missing; these
countries/economies have untrustworthy scale scores for the corresponding ISCED level.
5. This participating country/economy was excluded from this scale.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. These participating country/economy had a negative residual variance on item TC3G44E in the initial model,
resulting in the failure of the software to produce fit statistics for CFI, TLI and RMSEA. Several corrections
attempted to fix this issue, and eventually resulted in the residual variance of item TC3G44E being set to greater
than 0.01.
4. As the correction factor for this participating country/economy was negative, only the SRMR is reported.
5. Scale modelling failed in this participating country/economy.
6. This participating country/economy was excluded from this scale.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. As the correction factor for this participating country/economy was negative, only the SRMR is reported.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. This participating country/economy was excluded from this scale.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Table 11.128. Standardised factor loadings and unstandardised intercepts for the scale
T3PWLOAD
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
ones to exhibit factor loadings above 0.450, a pattern that suggests only a moderate
relationship with the latent construct.
T3PLEADS T3PLEADP
Participating countries/economies
Omega coefficient
ISCED level 2
Alberta (Canada) 0.929 0.796
Australia1 0.962 0.852
Austria 0.846 0.753
Belgium 0.778 0.803
Flemish Community (Belgium) 0.835 0.781
Brazil 0.859 0.884
Bulgaria 0.771 0.709
Ciudad Autónoma de Buenos Aires (Argentina) 0.828 0.815
Chile 0.867 0.887
Colombia 0.906 0.887
Croatia 0.823 0.887
Cyprus 0.899 0.874
Czech Republic 0.736 0.650
Denmark 0.910 0.663
England (United Kingdom) 0.943 0.726
Estonia 0.834 0.805
Finland 0.755 0.696
France 0.819 0.854
Georgia 0.780 0.887
Hungary 0.558 0.845
Iceland 0.841 0.701
Israel 0.912 0.778
Italy 0.884 0.681
Japan 0.746 0.599
Kazakhstan 0.702 0.752
Korea 0.916 0.828
T3PLEADS T3PLEADP
Participating countries/economies
Omega coefficient
Latvia 0.728 0.797
Lithuania 0.743 0.874
Malta 0.916 0.773
Mexico 0.835 0.774
Netherlands 0.845 0.755
New Zealand 0.884 0.778
Norway 0.814 0.769
Portugal 0.856 0.852
Romania 0.865 0.878
Russian Federation 0.750 0.927
Saudi Arabia 0.848 0.839
Shanghai (China) 0.922 0.899
Singapore 0.889 0.663
Slovak Republic 0.790 0.656
Slovenia 0.821 0.805
South Africa2, 3 0.819 -
Spain 0.817 0.908
Sweden 0.810 0.602
Chinese Taipei 0.814 0.701
Turkey 0.826 0.885
United Arab Emirates 0.901 0.889
United States 0.920 0.736
Viet Nam 0.733 0.796
ISCED level 1
Australia1 0.920 0.835
Flemish Community (Belgium) 0.870 0.561
Ciudad Autónoma de Buenos Aires (Argentina) 0.867 0.812
Denmark 0.880 0.621
England (United Kingdom) 0.914 0.814
France 0.895 0.669
Japan 0.714 0.712
Korea 0.978 0.835
Netherlands1 0.801 0.677
Spain 0.823 0.794
Sweden 0.776 0.834
Chinese Taipei 0.874 0.755
Turkey 0.876 0.922
United Arab Emirates 0.914 0.854
Viet Nam 0.746 0.740
ISCED level 3
Alberta (Canada) 0.856 0.785
Brazil 0.885 0.870
Croatia 0.906 0.845
Denmark 0.906 0.616
Portugal 0.854 0.738
Slovenia 0.778 0.759
Sweden 0.741 0.635
Chinese Taipei 0.803 0.856
Turkey 0.808 0.848
United Arab Emirates 0.914 0.863
Viet Nam 0.704 0.778
T3PLEADS T3PLEADP
Participating countries/economies
Omega coefficient
TALIS-PISA link
Australia 0.933 0.806
Ciudad Autónoma de Buenos Aires (Argentina) 0.812 0.712
Colombia 0.870 0.912
Czech Republic 0.815 0.846
Denmark 0.845 0.740
Georgia 0.841 0.882
Malta 0.897 0.787
Turkey 0.880 0.682
Viet Nam 0.884 0.960
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. This participating country/economy’s reliability coefficient estimation failed in the final scale model due to
a negative residual variance for one or more items that could not be corrected; this country/economy has
untrustworthy scale scores for the corresponding ISCED level.
Source: OECD, TALIS 2018 database.
School leadership
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. The poor fit of the model affected the TLI calculation, which is not reported.
Source: OECD, TALIS 2018 database.
Table 11.135. Unstandardised factor loadings for school leadership scales for all
participating countries/economies for all populations
T3PLEADS T3PLEADP
TC3G22D 0.476 TC3G26A 0.329
TC3G22E 0.597 TC3G26B 0.454
TC3G22F 0.499 TC3G26C 0.436
TC3G26D 0.282
TC3G26F 0.195
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
just below acceptable. Since the metric models were very near acceptable fit, each of these
cases was considered metric invariant. Lastly, Table 11.147 presents the results for the
scale T3PDELI, which reached configural invariance for all ISCED levels.
* This item was deleted from the scale and is not included in any of the results presented for this scale.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. This participating country/economy was excluded from the scale.
Source: OECD, TALIS 2018 database.
Academic pressure
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. Models in this county/economy included programmatic modifications.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. As the correction factor for this country/economy was negative, only the SRMR is reported.
Source: OECD, TALIS 2018 database.
Table 11.148. Unstandardised factor loadings for school climate scales for all participating
countries/economies for all populations
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. This participating country/economy was excluded from the scale.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
T3PORGIN
Participating countries/economies
Omega coefficient
ISCED level 2
Alberta (Canada) 0.830
Australia1 0.908
Austria 0.832
Belgium 0.792
Flemish Community (Belgium) 0.854
Brazil 0.891
T3PORGIN
Participating countries/economies
Omega coefficient
Bulgaria 0.885
Ciudad Autónoma de Buenos Aires (Argentina) 0.857
Chile 0.891
Colombia 0.904
Croatia 0.863
Cyprus 0.903
Czech Republic 0.850
Denmark 0.910
England (United Kingdom) 0.846
Estonia 0.801
Finland 0.812
France 0.878
Georgia 0.830
Hungary 0.906
Iceland 0.910
Israel 0.889
Italy 0.916
Japan 0.856
Kazakhstan 0.796
Korea 0.859
Latvia 0.857
Lithuania 0.906
Malta 0.865
Mexico 0.845
Netherlands 0.767
New Zealand 0.893
Norway 0.839
Portugal 0.846
Romania 0.874
Russian Federation 0.852
Saudi Arabia 0.835
Shanghai (China) 0.876
Singapore 0.841
Slovak Republic 0.941
Slovenia 0.642
South Africa2 0.852
Spain 0.823
Sweden 0.852
Chinese Taipei 0.976
Turkey 0.891
United Arab Emirates 0.916
United States 0.874
Viet Nam 0.578
ISCED level 1
Australia1 0.865
Flemish Community (Belgium) 0.785
Ciudad Autónoma de Buenos Aires (Argentina) 0.872
Denmark 0.937
England (United Kingdom) 0.920
T3PORGIN
Participating countries/economies
Omega coefficient
France 0.874
Japan 0.780
Korea 0.935
Netherlands1 0.850
Spain 0.906
Sweden 0.914
Chinese Taipei 0.964
Turkey 0.908
United Arab Emirates 0.906
Viet Nam 0.746
ISCED level 3
Alberta (Canada) 0.895
Brazil 0.846
Croatia 0.906
Denmark 0.931
Portugal 0.755
Slovenia 0.828
Sweden 0.843
Chinese Taipei 0.904
Turkey 0.887
United Arab Emirates 0.916
Viet Nam 0.640
TALIS-PISA link
Australia 0.891
Ciudad Autónoma de Buenos Aires (Argentina) 0.914
Colombia 0.912
Czech Republic 0.794
Denmark 0.924
Georgia 0.834
Malta 0.857
Turkey 0.899
Viet Nam 0.882
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Organisational innovativeness
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. As the correction factor for this participating country/economy was negative, only the SRMR is reported.
Source: OECD, TALIS 2018 database.
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
ISCED level 2 population for item TC3G40A. In summary, the relationship between the
items and the latent construct is strong in nearly all populations.
Table 11.159. Omega coefficient for the equity and diversity scale
T3PDIVB
Participating countries/economies
Omega coefficient
ISCED level 2
Alberta (Canada) 0.974
Australia1 0.945
Austria 0.872
Belgium 0.893
Flemish Community (Belgium) 0.899
Brazil 0.941
Bulgaria 0.910
Ciudad Autónoma de Buenos Aires (Argentina) 0.924
Chile 0.964
Colombia 0.920
Croatia 0.929
Cyprus 0.895
Czech Republic 0.850
Denmark 0.939
England (United Kingdom) 0.841
Estonia 0.870
Finland 0.901
France 0.918
Georgia 0.869
Hungary 0.922
Iceland 0.922
Israel 0.904
Italy 0.904
Japan 0.891
Kazakhstan 0.857
Korea 0.955
Latvia 0.933
Lithuania 0.845
Malta 0.937
Mexico 0.872
Netherlands 0.893
T3PDIVB
Participating countries/economies
Omega coefficient
New Zealand 0.968
Norway 0.904
Portugal 0.901
Romania 0.812
Russian Federation 0.699
Saudi Arabia 0.904
Shanghai (China) 0.941
Singapore 0.773
Slovak Republic 0.867
Slovenia 0.870
South Africa2 0.874
Spain 0.912
Sweden 0.865
Chinese Taipei 0.960
Turkey 0.958
United Arab Emirates 0.903
United States 0.970
Viet Nam 0.906
ISCED level 1
Australia1 0.955
Flemish Community (Belgium) 0.880
Ciudad Autónoma de Buenos Aires (Argentina) 0.939
Denmark 0.910
England (United Kingdom) 0.839
France 0.929
Japan 0.903
Korea 0.904
Netherlands1 0.912
Spain 0.924
Sweden 0.889
Chinese Taipei 0.941
Turkey 0.953
United Arab Emirates 0.912
Viet Nam 0.857
ISCED level 3
Alberta (Canada) 0.924
Brazil 0.945
Croatia 0.867
Denmark 0.925
Portugal 0.885
Slovenia 0.832
Sweden 0.887
Chinese Taipei 0.968
Turkey 0.941
United Arab Emirates 0.906
Viet Nam 0.852
TALIS-PISA link
Australia 0.943
Ciudad Autónoma de Buenos Aires (Argentina) 0.924
T3PDIVB
Participating countries/economies
Omega coefficient
Colombia 0.906
Czech Republic 0.812
Denmark 0.939
Georgia 0.904
Malta 0.933
Turkey 0.865
Viet Nam 0.884
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Diversity beliefs
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
3. As the correction factor for this participating country/economy was negative, only the SRMR is reported.
Source: OECD, TALIS 2018 database.
Table 11.162. Unstandardised factor loadings and intercepts for scale T3DIVB for all
participating countries/economies for all populations
T3PDIVB
Unstandardised factor loadings Unstandardised intercepts
TC3G40A 0.482 3.471
TC3G40B 0.500 3.486
TC3G40C 0.525 3.567
TC3G40D 0.526 3.546
1. Data from the participating country/economy was rated as insufficient during the adjudication process.
2. Participating country/economy with late data collection.
Source: OECD, TALIS 2018 database.
Distributed leadership
TC3G26: How strongly do you agree or disagree with these statements as applied to this school?
Response options: “Strongly disagree” (1), “Disagree” (2), “Agree” (3), “Strongly agree” (4).
TC3G26A This school provides staff with opportunities to actively participate in school decisions
TC3G26B This school provides parents or guardians with opportunities to actively participate in school decisions
TC3G26C This school provides students with opportunities to actively participate in school decisions
Equity beliefs
TC3G41: In your view, approximately how many teachers in this school would agree with the following statements?
Response options: “None or almost none” (1), “Some” (2), “Many” (3), “All or almost all” (4).
TC3G41A Schools should encourage students from different socio-economic backgrounds to work together
References
Beauducel, A., C. Harms and N. Hilger (2016), “Reliability estimates for three factor score [8]
estimators”, International Journal of Statistics and Probability, Vol. 5/6, pp. 94-107,
http://dx.doi.org/10.5539/ijsp.v5n6p94.
Brown, T. (2015), Confirmatory Factor Analysis for Applied Research: Second Edition, The [15]
Guilford Press, New York, NY.
Brown, T. (2006), Confirmatory Factor Analysis for Applied Research, Guilford Press, New [14]
York.
Chen, F. (2007), “Sensitivity of goodness of fit indexes to lack of measurement invariance”, [16]
Structural Equation Modeling, Vol. 14/3, pp. 464-504.
Chen, P., T. Cleary and A. Lui (2014), “Examining parents’ ratings of middle-school students’ [38]
academic self-regulation using principal axis factoring analysis”, School Psychology
Quarterly, Vol. 30/3, pp. 385-397, http://dx.doi.org/10.1037/spq0000098.
Cheung, G. (1999), “Testing Factorial Invariance across Groups: A reconceptualization and [23]
proposed new method”, Journal of Management, Vol. 25/1, pp. 1-27.
Cheung, G. and R. Rensvold (2002), “Evaluating goodness-of-fit indexes for testing [4]
measurement invariance”, Structural Equation Modeling, Vol. 9/2, pp. 233-255,
https://doi.org/10.1207/S15328007SEM0902_5.
Davidov, E. (2008), “A cross-country and cross-time comparison of the Human Values [24]
Measurements with the second round of the European Social Survey”, Survey Research
Methods, Vol. 2/1, pp. 33-46.
Davidov, E. et al. (2014), “Measurement equivalence in cross-national research”, Annual Review [27]
of Sociology, Vol. 40, pp. 55-75.
Dempster, A., N. Laird and D. Rubin (1977), “Maximum likelihood from incomplete data via the [34]
EM algorithm”, Journal of the Royal Statistical Society: Series B (Methodological),
Vol. 39/1, pp. 1-22, https://doi.org/10.1111/j.2517-6161.1977.tb01600.x.
DeVellis, R. (2003), Scale Development: Theory and Applications, Sage, Los Angeles, CA. [17]
Gonzalez, E. (2012), “Rescaling sampling weights and selecting mini-samples from large-scale [37]
assessment databases.”, IERI Monograph Series, Vol. 5, pp. 117-134.
Hansen, Y., M. Rosen and J. Gustavson (2006), “Measures of self-reported reading resources, [28]
attitudes and activities based on Latent Variable Modelling”, International Journal of
Research & Method in Education, Vol. 29/2, pp. 221-237.
He, Q. (2010), Estimating the Reliability of Composite Scores, Ofqual/10/4703, Office of [9]
Qualifications and Examinations Regulation, Coventry, https://dera.ioe.ac.uk/1060/1/2010-
02-01-composite-reliability.pdf.
Heywood, H. (1931), “On finite sequences of real numbers”, Proceedings of the Royal Society of [36]
London. Series A, Containing Papers of Mathematical and Physical Character, Vol. 134,
pp. 486-501.
Hoyle, R. (2014), Handbook of Structural Equation Modeling, The Guilford Press, New York, [18]
NY.
Hu, L. and P. Bentler (1999), “Cutoff criteria for fit indexes in covariance structure analysis: [11]
Conventional criteria versus new alternatives”, Structural Equation Modeling, Vol. 6/1, pp. 1-
55, https://doi.org/10.1080/10705519909540118.
Lüdtke, O. et al. (2007), “Umgang mit fehlenden Werten in der psychologischen Forschung”, [33]
Psychologische Rundschau, Vol. 58/2, pp. 103-117, http://dx.doi.org/10.1026/0033-
3042.58.2.103.
Muthén, B. (1977), Some results on using summed raw scores and factor scores from [32]
dichotomous items in the estimation of structural equation models, Unpublished Technical
Report, University of Uppsala, Sweden.
Muthén, L. and B. Muthén (1998-2017), Mplus User’s Guide, Muthén & Muthén, Los Angeles, [30]
CA.
OECD (2019), TALIS 2018 Results (Volume I): Teachers and School Leaders as Lifelong [1]
Learners, TALIS, OECD Publishing, Paris, https://dx.doi.org/10.1787/1d0bc92a-en.
OECD (2010), TALIS 2008 Technical Report, TALIS, OECD Publishing, Paris, [12]
https://dx.doi.org/10.1787/9789264079861-en.
Schafer, J. and J. Graham (2002), “Missing data: our view of the state of the art”, Psychological [35]
Methods, Vol. 7/2, pp. 147-177, http://dx.doi.org/10.1037/1082-989X.7.2.147.
Schermelleh-Engel, K., H. Moosbrugger and H. Müller (2003), “Evaluating the fit of structural [19]
equation models: Tests of significance and descriptive goodness-of-fit measures”, Methods of
Psychological Research Online, Vol. 8/2, pp. 23-74.
Schreiber, J. et al. (2006), “Reporting Structural Equation Modeling and Confirmatory Factor [13]
Analysis results: A review”, The Journal of Educational Research, Vol. 99/6, pp. 323-338.
Skrondal, A. and P. Laake (2001), “Regression among factor scores”, Psychometrika, Vol. 66, [31]
pp. 563-575, https://doi.org/10.1007/BF02296196.
Steiger, J. (1990), “Structural model evaluation and modification: An interval estimation [20]
approach”, Multivariate Behavioral Research, Vol. 25/2, pp. 173-180.
Van de Vijver, F. et al. (2019), “Invariance analyses in large-scale studies”, OECD Education [26]
Working Papers, No. 201, OECD Publishing, Paris, https://dx.doi.org/10.1787/254738dd-en.
Vandenberg, R. and C. Lance (2000), “A review and synthesis of the measurement invariance [5]
literature: Suggestions, practices, and recommendations for organizational research”,
Organizational Research Methods, Vol. 3/1, pp. 4-70.
Viladrich, C., A. Angulo-Brunet and E. Doval (2017), “A journey around alpha and omega to [6]
estimate internal consistency reliability”, Anales de Psicología, Vol. 33/3, pp. 755-782.
Yu, C. (2002), Evaluating Cutoff Criteria of Model Fit Indices for Latent Variable Models with [21]
Binary and Continuous Outcomes, unpublished doctoral dissertation, University of
California, Los Angeles.
Zhang, Z. and K. Yuan (2016), “Robust coefficients alpha and omega and confidence intervals [7]
with outlying observations and missing data: Methods and software”, Educational and
Psychological Measurement, Vol. 76/3, pp. 387-411,
https://doi.org/10.1177/0013164415594658.
Notes
1
The question specifies to consider only teachers whose main activity is the provision of instruction
to students.
2
Personnel for pedagogical support include: (a) teacher aides or other non-teaching professionals
who provide instruction or support teachers in providing instruction; (b) professional
curricular/instructional specialists; and (c) educational media specialists, psychologists and nurses.
3
School administrative or management personnel include principals, assistant principals and other
management staff whose main activity is administration or management.
4
These indices are composed of items assessing decision making process and numbers of resources,
which are not latent constructs and, therefore, not appropriate for CFA.
5
The models used in the scaling procedure account for random measurement error, but also
recognise that variables are measured with certain but not complete precision.
6
A pooled sample is a dataset where all the countries/economies from ISCED level 2 are put together
and analysed as one group. All factor loadings and item intercepts are freely estimated and the latent
variances is fixed to one.
7
Constrained models were used for “insufficient” populations (see Chapter 10 for more details),
that is, countries/economies with late data submission and TALIS-PISA link populations. These
models are based on the parameters of the final scale models (see the section “Scale score
estimation” for further details).
8
For practical reasons, the two most recent versions of Mplus were used (versions 8 and 7.3).
9
The evaluation procedures in the field trial were based on continuous and categorical models so as
to assess the comparability of the results. This procedure was chosen because the majority of TALIS
items are ordinal. However, in both previous cycles (TALIS 2008 and 2013), linear models were
used to estimate complex indices based on ordinal scale items. In TALIS 2018, both models were
used to evaluate the scales, followed by a comparison of their results. The results were very similar
although there were minor differences in terms of the performance and the measurement invariance
level of the scale. Practical challenges related to a possible change in approach (i.e. from continuous
CFA to categorical CFA modelling) between TALIS cycles resulted in the decision to implement
the linear measurement model for TALIS 2018 scaling. This approach is supported by simulation
studies (Van de Vijver et al., 2019[26]).
10
Weighted least square parameter estimates using a diagonal weight matrix with standard errors
and mean- and variance-adjusted chi-square test statistic that use a full weight matrix (Muthén and
Muthén, 1998-2017[30])
11
So that each participating country/economy contributes equally to parameter estimation (that is, a
population should not have more or less influence on parameter estimates due to its size)
countries’/economies’ sampling weights were rescaled to sum up to the same number of teachers
and principals. This way, each participating country/economy makes an equal contribution to the
estimates (Gonzalez, 2012[37]). The teacher and principal weights for each participating
country/economy in each ISCED level were rescaled so that the total sum of the weights (i.e. number
of teachers and principals in the population) was equal to 3,000 and 200 respectively. This way,
participating countries/economies make an equal contribution in parameter estimation regardless of
the sizes of their samples or populations. The SPSS macro provided by Gonzalez (Gonzalez,
2012[37]) was used for rescaling.
12
A more detailed examination of the model parameters (factor loadings, residual variances) was
part of further model analyses.
13
Principal axis factoring (PAF) is a form of EFA that is commonly adopted to examine the internal
factor structure of constructs. Unlike principal component Analysis (PCA), which is a linear
combination of variables, PAF is a measurement model of latent constructs. Oblimin rotation was
chosen over Varimax rotation due to the assumption that the extracted factors within the constructs
in TALIS 2018 are correlated with one another (Brown, 2006[14]; Chen, Cleary and Lui, 2014[38]).
14
Participating countries/economies that did not meet the Technical standards, participating
countries/economies with late data collection, and TALIS-PISA link populations were not included
in this stage of the analysis. For a detailed explanation see the sub-sectionDescription of scales and
their parameters and Table 11.7.
15
Thirteen countries/economies participated at the ISCED level 1, 48 countries/economies at ISCED
level 2, 11 countries/economies at ISCED level 3.
16
The measurement parameters describe measurement characteristics of observed variables (items).
Factor loadings indicate the strength of the relationship between the item and the latent factor,
intercepts indicate the predicted values for an item when the value of the latent factor is zero, and
residual variances presents the portion of item variance that is not attributable to variance in the
latent factor.
17
For these models, a minimum number of parameters are constrained for reasons of model
identification. In the analysis, the latent variances was set to one for each group allowing for free
estimation of factor loadings and intercepts.
18
In strict invariance, not only the factor loadings and item intercepts but also the residual variances
of the items have to be equal across groups. This requirement means that that the portion of the item
variance not attributable to variance in the latent construct is the same across groups. However, this
assumption is very hard to meet in practice especially in large-scale assessments where many groups
are compared. In line with previous TALIS cycles, the current cycle did not test these models.
19
To be specific, for a certain ISCED level, p CFA models were created, where p is the number of
countries/economies within that ISCED level.
20
During TALIS 2018, one, two or three models were estimated for each participating
country/economy, in accordance with the number of populations that each participating
country/economy participated in for the current cycle. For example, some participating
countries/economies participated in each of the three ISCED levels, meaning three models were
estimated. Other countries/economies participated in ISCED level 2 and one other, either ISCED
level 1 or ISCED level 3, meaning two models (either ISCED level 2 and ISCED level 1 or ISCED
level 2 and ISCED level 3) were estimated.
21
Configural, metric, and scalar models are in essence nested models; the scalar model is nested in
the metric model, and the metric model is nested in the configural model. Therefore, the χ2 difference
test can be adopted to evaluate which model fits the data best. If the χ2 difference value is significant,
the less restrictive model (the model with more freely estimated parameters) fits the data better than
the nested more restrictive model (the model with fixed/constrained parameters). If the χ2 difference
value is not significant, then both models fit the data equally well. However, because χ2 is sensitive
to sample size, which were particularly large for TALIS, the change of the model fit indices (e.g.,
CFI, TLI, RMSEA, SRMR/WRMR) were used to evaluate the measurement invariance of each
scale.
22
For scale measured with categorical variables, the standardised factor loadings come from the
STDYX standardisation of Mplus, while for scales measured with binary items they come from the
STDY standardisation (Muthén and Muthén, 1998-2017, pp. 799-802[30]).
23
The mean score does not always rise and fall in this way. Certain aggregate scores are observed
much less than others. In this example, with the scale TPERUT, aggregate scores of 1.33, 1.67, 2.33,
2.67, 3.33 and 3.67 are less common resulting in a smaller number of observations for these values.
Therefore, the variance is not very large, an outcome that can result in a mean that is lower or greater
than expected. However, the correlation between the simple average of items scores and factor
scores is still generally strong.
24
Due to the complexity of the final models, which took into account measurement invariance
testing results both cross-country within each ISCED level, and cross-ISCED levels with each
participating country/economy participating in more than one population sample, and the structure
of the large-scale data, multidimensional models needed computing power beyond the limit of
available resources
25
The TALIS 2013 technical report uses the term “calibration sample” to refer to the ISCED level
2 population, which is a different use of the term from the TALIS 2008 technical report, in which it
referred to a random sample of each participating country/economy so that each contributed to the
model parameter estimates equally.
26
For example, in relation to teacher job satisfaction, a positive response of “strongly agree” to the
item “The advantages of being a teacher clearly outweigh the disadvantages” is also positive for the
latent construct of job satisfaction. However, a positive response of “strongly agree” to the item “I
regret that I decided to become a teacher” is negatively related to the latent construct, which means
this item would be reverse coded; therefore, a response of “strongly disagree” would now be coded
(i.e. assigned a higher integer) to align with the positive association with the latent construct for this
particular item.
27
The CFA on a pooled sample during the evaluation process was conducted to gain a general
overview of the scale’s performance. The results from this analysis are therefore not presented in
this technical report. In addition, model parameters were used to fix the results from the final scale
modelling for populations that did not meet the technical standards and for those participating
countries/economies submitting data late (see Chapter 10 for more detail). These participating
countries/economies were not considered during model analysis at the country/economy level.
28
In models with three items, the number of the information in the variance covariance matrix equals
the number of parameters being estimated. The fact that the number of degrees of freedom equals
zero (𝑑𝑓 = 0) provides a unique solution to the parameter estimation and means that the models are
just identified. Consequently, models based on three items could not be evaluated with respect to
their fit to the empirical data (because the model fit indices suggest a perfect model fit).
29
The composite scale is evident in the variable labels in the database referred to as “overall”.
30
The complexity of the final mode meant the multidimensional scales could not be processed
because of the resulting lack of computation power.
31
Their data were rated as “insufficient” during the adjudication process. For details, see Chapter 10.
32
The TALIS-PISA link populations were not included into the parameter estimation as it is not part
of the TALIS target populations.
33
Because the models in these populations were specified by fixing the parameters (factor loadings
and intercepts) to be equal to the international parameters, the number of the degrees of freedom
was positive (𝑑𝑓 > 0) and the model fit could therefore be estimated.
34
The weights for the teacher data were slightly altered late in the analysis process. To check the
impact of changes on the invariance testing results, the TALIS Consortium performed invariance
tests for a few scales to ensure that decisions did not change. After confirming that the change only
minimally affected the numeric results, did not affect the model evaluation and invariance testing
results and did not affect the decisions made from such results, the consortium moved on to the final
scale modelling (computing the scale scores). Due to the weights being altered late in the modelling
some of the results are reported with the use of preliminary weights (before the minor alteration), as
indicated by this endnote.
This chapter describes the table production process, from developing the table shells up to
the verification of the table results. It provides an overview of conducted analyses, applied
quality rules, software used and the different parties involved in ensuring high-quality
results.
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
12.1. Introduction
The process and procedures presented in this report and, in particular, this chapter, relate
to the tables of results produced by the International Association for the Evaluation of
Educational Achievement (IEA) Hamburg and the OECD for Volume I of the TALIS 2018
international report (OECD, 2019[1]). Preparation of the tables consisted of two major steps:
(1) the development, review and revision of table shells; and (2) data analysis, table
production and verification. After providing a brief description of the first step, this chapter
focuses on the second step. The chapter also provides an overview of the procedures and
methods applied to estimate population parameters and uses selected but shortened
versions1 of the tables from the report to highlight relevant statistical issues. It furthermore
provides insights into the production and verification of regression tables.
12.2. Responsibilities
The OECD and the TALIS 2018 International Consortium, specifically the IEA and the
Australian Council for Education Research (ACER), shared the development and
production of the tables for the TALIS 2018 international report (OECD, 2019[1]). The
OECD developed the table shells and determined the content and format of the tables on
the basis of the analysis plan produced by the TALIS Consortium, as well as the reporting
plan developed by the OECD and agreed to by the participating countries/economies. The
table shells were then reviewed by the TALIS Consortium for statistical and substantial
soundness and revised by the OECD over multiple rounds.
In a subsequent step, the TALIS Consortium’s IEA Hamburg team conducted the analyses,
populated around 190 tables in collaboration with the OECD and verified/validated
estimates in collaboration with ACER and the OECD. The IEA team members discussed
methodological issues relating to these tables amongst themselves and with the OECD.
Both the IEA team and the OECD considered advice from the Technical Advisory Group
(TAG). The OECD was also responsible for designing and populating an additional 45
tables showing the results of regression analyses. The OECD furthermore noted the
consortium’s recommendations and shared them in the TALIS 2018 analysis plan and
consulted with the TAG. The countries/economies participating in TALIS 2018
participated in the table development and review process via the OECD Secretariat.
The process of populating table shells consisted of two major steps. First, the IEA IDB
Analyzer was used to estimate, for each table, all parameters (percentages, means, etc.) and
their respective standard errors.2 Second, R packages3 were used to transfer estimates
resulting from the first step to the appropriate table shell and cell; estimates for the EU total,
the OECD and TALIS averages were computed (see the section OECD average, TALIS
average and EU total below); and the supplementary statistical analyses needed for flagging
estimates4 and quality control (such as the analysis of non-response at item level and the
verification of a minimum sample size) were conducted. The codebook provided in
Annex J of this report includes information on: (1) the variables created and delivered in
the public TALIS 2018 dataset; and (2) the advanced recoded variables generated during
table production.
All tables presented in the TALIS 2018 initial report were structured in a manner similar
to the example shown in Table 12.1, and each accorded with these elements and principles:
1. A parameter always appears with its standard errors (SE) (Items 1 and 2,
Table 12.1).
2. The following averages appear (see Items 4 to 6 in Table 12.1, formula presented
in the section on OECD average, TALIS average and EU total) in all tables except
those tables displaying change over time, tables based on the ISCED levels 1 and 3
populations and selected tables featuring data from survey questions that were
co-ordinated national options:
OECD average-30, OECD average-31: average of the OECD countries/
economies participating in TALIS 2018
EU total-23: total of the 23 EU countries/economies participating in TALIS
2018
TALIS average-47, TALIS average-48: average of all the countries/economies
participating in TALIS 2018.
3. Country/economy-specific considerations:
a. “Flemish Comm.” (i.e. Flemish Community): This name, written in italics and
with an indentation, indicates separate adjudication. However, note that data
collected within the Flemish Community of Belgium (Item 3) also contributed
to the estimates for Belgium.
b. Data from countries/economies not meeting the requirements determined in the
technical standards (see Chapter 10 of this technical report): These are reported
below the main body of the table. The affected data did not contribute to the
above-cited averages (see Items 4 to 7 in Table 12.1).
Percentage of principals who participated in the following professional development activities in the previous 12 months:
Courses/seminars about subject matter, teaching methods or pedagogical topics
% (1) SE (2)
Alberta (Canada) 62.2 (11.0)
Austria 82.7 (3.1)
Belgium 78.9 (3.0)
Flemish Comm. (Belgium) (3) 85.8 (3.2)
… … …
… … …
… … …
Turkey 46.9 (4.2)
United Arab Emirates 70.0 (2.0)
United States 73.7 (7.5)
Viet Nam 87.9 (2.6)
OECD average-30 (4) 70.5 (0.8)
EU total-23 (5) 66.0 (1.1)
TALIS average-47 (6) 73.1 (0.6)
Australia (7) 79.7 (4.4)
One of TALIS’s main goals is to generate reliable, valid and comparable population
estimates based on sample data. All parameters presented in the TALIS 2018 international
report are therefore weighted (see, for example, the percentages in Table 12.1). The data
provided by school principals that contributed to the school-based estimates were weighted
by school weights (variable name: SCHWGT), while the results arising out of either teacher
data or combined teacher and school principal data (i.e. school/principal information
merged with teacher records) were weighted by teacher weights (variable name:
TCHWGT).
Because all estimates in the TALIS 2018 international report are based on sample data,
they could only be estimated with a degree of uncertainty. Thus, results from analyses of
these data and the information on the precision of the population estimates must be reported
together. In the TALIS tables, the degree of uncertainty of an estimate is reflected by its
standard error (SE; Item 2 in Table 12.1) and this has the same metric as the estimate.
Fay’s variant of the balanced repeated replication (BRR) technique was used to estimate
the standard errors during the weighting stage. BRR estimates of sampling error can be
computed with the IEA IDB Analyzer.
Chapter 9 of this report gives more information on computing sampling weights and about
the BRR technique. It also describes how to obtain standard errors for differences of
estimates obtained from different samples or from the same sample. Chapter 3 of the TALIS
2018 User Guide to the International Database (OECD, forthcoming[2]) provides details
on using weights for data analysis.
The formula used to calculate the OECD and TALIS averages was:
𝐶
1
𝜃̂̅ = ∑ 𝜃̂𝑐
𝐶
𝑐=1
where C is the number of countries contributing to the average 𝜃̂̅ , 𝑐 is an index that runs
from 1 to C, and 𝜃̂𝑐 is the estimate for country 𝑐.
The formula used to calculate the standard errors for the OECD and the TALIS averages
was:
𝐶
̂ ∑𝐶𝑐=1 𝑠𝑒( 𝜃̂𝑐 )2 ∑𝐶𝑐=1 𝑉̂𝐹𝑎𝑦 (𝜃̂𝑐 ) 1
̅
𝑠𝑒 (𝜃 ) = √ = √ = √∑ 𝑉̂𝐹𝑎𝑦 (𝜃̂𝑐 ).
𝐶2 𝐶2 𝐶
𝑐=1
Here, C is the number of countries contributing to the average 𝜃̂̅ , 𝑐 is an index that runs
from 1 to C, 𝜃̂𝑐 is the estimate for country 𝑐, and 𝑉̂𝐹𝑎𝑦 (𝜃̂𝑐 ) is the Fay’s BRR estimate for
the variance of that estimate. According to this formula for statistically independent
samples, the standard error of the average 𝑠𝑒 (𝜃̂̅ ) is the square root of the sum of the
squared standard errors divided by the squared number of countries.
The formula used to calculate the estimate for the EU total (Table 12.1, Item 5) was:
∑𝐶 ̂ ̂
𝑐=1 𝑁𝑐 𝜃𝑐
𝐸𝑈 𝑡𝑜𝑡𝑎𝑙 = ∑𝐶 ̂
,
𝑐=1 𝑁𝑐
̂𝑐 is the estimated population size for country 𝑐, that is, the sum of the appropriate
where 𝑁
weights (e.g. for principal-level analyses, the sum of the school weights of country 𝑐), C is
the number of countries contributing to the estimate of the 𝐸𝑈 𝑡𝑜𝑡𝑎𝑙, and 𝜃̂𝑐 is the
parameter of interest (i.e. a country-specific average of the analysis variable) for country 𝑐.
The formula used to calculate the standard error of the EU total (see Table 12.1, Item 5, SE
column) was:
∑𝐶 𝑁 ̂𝑐2 𝑠𝑒(𝜃̂𝑐 )2
𝑠𝑒𝐸𝑈 𝑡𝑜𝑡𝑎𝑙 √ 𝑐=1
̂𝑐 )2
(∑𝐶𝑐=1 𝑁
where 𝑁 ̂𝑐 is the estimated population size for country 𝑐, 𝐶 is the number of countries
contributing to the estimate 𝑠𝑒𝐸𝑈 𝑡𝑜𝑡𝑎𝑙 , and 𝜃̂𝑐 is the parameter of interest (i.e. a country-
specific average of the analysis variable).
While all countries were weighted to equally contribute to the OECD and TALIS averages,
this was not the case for the EU total. Here, each country/economy contributed according
to its estimated population size, meaning that countries/economies with a large estimated
population size (i.e. a high number of teachers or principals) contributed more to the EU
total than did countries/economies with a small population size (i.e. with a low number of
teachers or principals). Thus, the “EU total” is an “average European teacher”, whereas the
OECD or TALIS average is an “average country/economy”.
The Flemish Community of Belgium (see Table 12.1, Item 3) was excluded from
calculations of the OECD average, TALIS average and EU total, as the population was
already included in the estimates for Belgium. Countries/economies for which estimates
are flagged in the table (e.g. because the respective question was not administered) were
not included in the calculation of the OECD average, TALIS average and EU total.
The IEA IDB Analyzer was used to estimate the percentiles. This procedure had two steps.
Step 1 involved sorting the values and producing a vector of accumulated weighted
frequencies. During Step 2, the first value to exceed the percentile threshold (25th, 50th
and 75th) became the respective percentile value.
12.7. Use of weights in tables featuring analyses of teachers’ and principals’ data
In tables including results derived from both principal and teacher data, the column
displaying results from the principal questionnaire has to be interpreted with caution
because of the estimation algorithm, which required the principal data to be merged with
the teacher data, and the teacher weights then used to conduct the analyses. The principal
data in these tables must, therefore, be viewed as a feature of teachers rather than of the
actual principal population.
Box 12.1. Interpretation of results derived from analysis of teachers’ and principals’ data
As an example of how to interpret the results derived from jointly analysing principals’
data with teachers’ data, consider the results displayed in the column labelled “principal”
in Table 12.2. The estimates in this column need to be interpreted as the percentages of
teachers whose principal said that his or her school was “supporting activities or
organisations encouraging students’ expression of diverse ethnic and cultural identities”. It
would, therefore, be incorrect to state that a specific percentage of principals said their
schools were “supporting activities or organisations encouraging students’ expression of
diverse ethnic and cultural identities”.
Percentage of teachers working in a school with diverse ethnic and cultural student background where the following diversity-
related practices are implemented:
Supporting activities or organisations encouraging students’ expression of diverse ethnic and cultural identities
According to teachers According to principals
% SE % SE
Alberta (Canada) 75.2 (2.2) 83.9 (5.0)
Austria 50.6 (1.5) 64.8 (3.5)
Belgium 52.1 (1.3) 59.3 (3.1)
Flemish Comm. (Belgium) 47.5 (1.7) 48.6 (4.3)
… … … … …
… … … … …
… … … … …
Turkey 40.8 (1.3) 48.5 (0.9)
United Arab Emirates 88.1 (0.6) 94.4 (0.2)
United States 72.8 (2.5) 90.2 (2.7)
Viet Nam 94.8 (1.2) 93.5 (3.0)
OECD average-30 54.2 (0.4) 61.3 (0.8)
EU total-23 57.0 (0.5) 67.0 (1.2)
TALIS average-47 62.7 (0.3) 69.9 (0.6)
Another important TALIS goal is to report change over time. As the third cycle of the
TALIS survey, TALIS 2018 enables data users to look at changes over a ten-year period.
However, this type of analysis poses particular challenges that data users need to consider
when interpreting comparisons across cycles. First, not all countries/economies
participated in all cycles. Second, study questionnaires underwent some modifications,
which means that not all questions remained the same across the three measurement points.
A third consideration is that the context of teaching and learning may have changed
considerably across the ten years in terms of student intake or general societal trends, for
example. This section describes the measures taken to address these challenges.
Samples in the tables displaying different TALIS cycles were adjusted to represent the
same populations across all cycles, thus ensuring comparability over time. This adjustment
was implemented by using the filter variable TALIS13POP to exclude affected cases from
analyses based on the TALIS 2018 dataset – see also OECD (forthcoming[2]). Chapter 9 of
this report provides the formula for estimating the standard error of the estimated difference
between cycles (i.e. the standard error for the difference between the estimates for two
countries/economies).
A change in the defined target populations across the TALIS cycles also needs to be
considered when comparing findings across them. Whereas in 2008, teachers of special
needs children in regular schools were not part of the target population definition and were
therefore not covered (OECD, 2010, p. 56[3]), these teachers were included in the TALIS
2013 and TALIS 2018 surveys (see Chapter 5). Comparisons of the estimates from TALIS
2008 and the other two cycles (2013 and 2008) therefore need to be made with caution.
Note also that teachers in schools exclusively directed towards teaching students with
special educational needs were excluded from all three cycles.
The OECD was responsible for conducting the regression analyses and for producing the
subsequent 45 tables presenting the regression results. The table shells for these analyses
were designed with the aim of ensuring that the analyses aligned with the plan for reporting
TALIS results.
The regression analyses, which were conducted separately for each participating
country/economy, explored the relationships between different variables. Linear regression
was used in those cases where the dependent (or outcome) variable was considered
continuous. Binary logistic regression was employed if the dependent (or outcome) variable
was a binary variable.
Selection of the independent (or control) variables included in each regression model was
based on theoretical reasoning and, preferably, limited to the most objective measures or
those measures that do not change over time. Controls for teacher characteristics included
teacher’s gender, age, employment status (full-time/part-time) and years of teaching
experience. Controls for class characteristics included variables of classroom composition
(share of students whose first language differed from the language of instruction, low
achievers academically, students with special needs, students with behavioural problems,
students from socio-economically disadvantaged homes, academically gifted students,
immigrant students or students with an immigrant background, refugee students) and class
size.
In the case of the multiple linear regression models, the models’ explanatory power is also
highlighted by the R-squared (R²), which represents the proportion of the observed
variation in the dependent (or outcome) variable that can be explained by the independent
(or control, explanatory) variables.
To ensure the robustness of the regression models, the OECD team used a stepwise
approach to introduce the independent variables into the models. This approach required
each step of the model to be based on the same sample. Consequently, during analysis, the
restricted sample used for the different versions of the same model corresponded to the
sample of the most extended version of the model (i.e. the version with the maximum
number of independent variables). Thus, the restricted sample for each regression model
excluded those observations that had missing values for any of the independent variables.5
Some questions in the TALIS surveys are meant to be answered by a defined subgroup of
the surveys’ target populations. The respective subgroups are identified by their responses
to filter questions. Consider, for example, Question 37 in the TALIS 2018 principal
questionnaire (variable TC3G37): “In this school, are the following policies and practices
in relation to diversity implemented?”. This question was to be answered only by principals
who gave the answer “Yes” to the preceding filter question (variable TC3G37): “Does this
school include students of more than one cultural or ethnic background?”. In this instance,
TC3G37 is a filter question and TC3G38 is the filter-dependent question.
Estimates involving filter-dependent questions were based only on those respondents who
were filtered in by the preceding corresponding filter question (information on this matter
can be found in the table header of each table that provides the selection criteria for the
filter). The IEA and OECD team implemented this rule by treating cases not meeting the
filter criteria (i.e. observations with logically not applicable codes in the dataset) as missing
values that needed to be ignored in the analyses (more information about the missing codes
can be found in Chapter 8 of this report). If a participant did not answer a filter question,
yet answered a subsequent filter-dependent question, the team included the response to the
filter dependent question in the analyses.
In specific cases, subgroups were defined not only by filter variables but also by other
criteria for interpretative reasons. Information on these criteria is provided in the footnotes
below the tables of the TALIS 2018 international report, TALIS 2018 Results (Volume I):
Teachers and School Leaders as Lifelong Learners (OECD, 2019[1]).
The annotation scheme shown in Box 12.2 for flagging empty cells in the tables of the
TALIS 2018 international report was developed by the TALIS International Consortium in
collaboration with the OECD and the TAG and it followed the approaches used for the
TALIS 2013 and PISA 2015 international reports.
a The question was not administered in the country/economy because it was optional
or was part of a questionnaire from a TALIS cycle the country/economy did not
participate in. Data are therefore deemed missing data.
c There were too few or no observations to provide reliable estimates and/or to ensure
the confidentiality of respondents, that is, there were fewer than
10 schools/principals and/or 30 teachers with valid data; and/or the item
non-response rate (i.e. ratio of missing or invalid responses to the number of
participants for whom the question was applicable) was above 50%.
m Data were collected but were subsequently, as part of the data-checking process,
removed for technical reasons (e.g. erroneous translation).
p Data were collected but not reported for technical reasons (e.g. low participation
rate) as part of the data adjudication process.
w Data were withdrawn or were not collected at the request of the country/economy
concerned.
The first data quality check verified the proportion of participants who answered a given
question (item non-response rule). The item non-response rule was implemented to ensure
that the number of participants who gave a valid answer to a question divided by the number
of participants to whom the question was applicable (refers to filter or CNO)6 was not lower
than 50%. The second data quality check ensured that an estimate was based on at least 30
teachers and 10 principals/schools (minimum sample rule). If a country/economy did not
meet one or both requirements, the analysis was not conducted for that country/economy,
as indicated by the symbol “c” in the relevant table. For example, in the TALIS dataset for
Alberta (Canada), no more than ten principals from privately managed schools gave a valid
answer to the question “Are you female or male?”, which accounts for the “c” for Alberta
(Canada) in the Table 12.3 column titled “privately managed schools”.
Estimated differences that tested significantly different from zero are highlighted in bold
in the tables. More information on statistical testing can be found in the TALIS 2018 and
TALIS Starting Strong Survey 2018 User Guide (OECD, forthcoming[2]).
To ensure high quality, the results presented in the tables were produced and released
during up to three review and revision rounds. During each release, the TALIS Consortium,
the OECD and the participating countries/economies verified the results for substantial
plausibility.
To validate and verify the statistical procedures applied by the TALIS Consortium
(implemented with the IEA IDB Analyzer and R), both ACER and the OECD reproduced
all estimates presented in the tables. ACER recalculated the table results using ACER-
developed tools for replicated analysis (Fay-BRR in this case), while the OECD relied on
the repest function in STATA.
Verification encompassed the following activities:
general plausibility checks
different reviewers verifying the variables, sub-settings and recodings several times
over
recalculation of randomly-selected results from each table (usually one or
two randomly-selected columns of a table)
recalculation of all OECD and TALIS averages and the EU totals
comparison of results against the results produced by ACER and the OECD.
References
OECD (2019), TALIS 2018 Results (Volume I): Teachers and School Leaders as Lifelong [1]
Learners, TALIS, OECD Publishing, Paris, https://dx.doi.org/10.1787/1d0bc92a-en.
OECD (2010), TALIS 2008 Technical Report, TALIS, OECD Publishing, Paris, [3]
https://dx.doi.org/10.1787/9789264079861-en.
OECD (forthcoming), TALIS 2018 and TALIS Starting Strong Survey 2018 User Guide, OECD, [2]
Paris.
Notes
1
Tables were shortened to highlight important statistical aspects and thus to increase readability.
All footnotes that were not important in the context of this chapter were deleted from the tables, but
can be found in the international report (OECD, 2019[1]). Order and names of countries in Chapter 12
tables match tables in the international report (OECD, 2019[1]) but differ slightly from other tables
in this technical report.
2
The IEA IDB Analyzer is a software application developed by the IEA to perform analysis with
data of international large-scale assessments in education. It provides, via a user interface, SPSS or
SAS syntax that can be run with the respective software (see www.iea.nl/data). Version 4.0.26 of
the IEA IDB Analyzer was used to produce the TALIS 2018 tables.
3
The following R packages were used: data.table, openxlsx. More information on R can be found
at https://www.r-project.org.
4
Box 12.2 of this chapter details the annotation rules relating to data quality requirements and their
associated flagging symbols.
5
Listwise deletion was implemented in the regression analysis.
6
CNO refers to co-ordinated national options or, in other words, to country-specific questions.
Core management
Steffen Knoll (International Study Co-Director Operations)
Ralph Carstens (International Study Co-Director Content)
Meeting organisation
Catherine Pfeifer (Meeting Co-ordinator)
Bettina Wietzorek (Meeting Co-ordinator and SharePoint Administrator)
ICT services
Malte Bahrenfuß (ICT Service Manager)
Jan Pohland (ICT Service Provider)
Consultants
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
Participating Population and Reasons for exclusions Number of Percentage Number of Percentage
country/economy coverage schools of total teachers of total
Australia Target Population 7 733 100 140 827 100
Exclusions: Small schools 1 106 14.3 2 289 1.63
Schools that are geographically remote 77 1 533 0.38
Small schools that are also 40 0.52 119 0.08
geographically remote
Survey Population 6 510 84.18 137 886 97.91
Flemish Target Population 2 238 100 .. ..
Community
(Belgium)
Exclusions: Schools with fewer than 20 pupils at 5 0.22 .. ..
ISCED level 1
Survey Population 2 233 99.78 .. ..
Ciudad Autónoma Target Population 878 100 .. ..
de Buenos Aires
(Argentina)
Survey Population 878 100 .. ..
Denmark Target Population 1 751 100 .. ..
Exclusions: Small schools (fewer than 40 students 57 3.26 .. ..
and generally fewer than 5 teachers)
Survey Population 1 694 96.74 .. ..
England Target Population 18 445 100 .. ..
(United Kingdom)
Exclusions: International schools 17 0.09 .. ..
Very small schools 258 1.4 .. ..
Schools proposed for closure 26 0.14 .. ..
Survey Population 18 144 98.37 .. ..
France Target Population 33 929 100 .. ..
Exclusions: Schools located in overseas French 105 0.31 .. ..
territories (TOM)
Schools located in La Réunion and 473 1.39 .. ..
Mayotte (Southern Hemisphere
calendar)
Private schools under different 425 1.25 .. ..
administration
Survey Population 32 926 97.04 .. ..
Japan Target Population 20 333 100 385 923 100
Survey Population 20 333 100 385 923 100
Korea Target Population 6 242 100 118 417 100
Exclusions: Schools with fewer than three teachers 120 1.92 237 0.2
at ISCED level 1
Schools that are geographically remote 303 4.85 1 809 1.53
Schools with fewer than three teachers 208 3.33 305 0.26
at ISCED level 1 and also
geographically remote
Survey Population 5 611 89.89 116 066 98.01
Participating Population and Reasons for exclusions Number of Percentage Number of Percentage
country/economy coverage schools of total teachers of total
Netherlands Target Population 6 429 100 111 157 100
Exclusions: Fewer than five teachers 95 1.48 339 0.3
Schools on wheels/boat schools 2 0.03 49 0.04
Survey Population 6 332 98.49 110 769 99.65
Spain Target Population 13 603 100 287 262 100
Exclusions: Small schools 328 2.41 800 0.28
Survey Population 13 275 97.59 286 462 99.72
Sweden Target Population 4 339 100 .. ..
Exclusions: Small schools (fewer than six students 67 1.54 .. ..
at ISCED level 1)
International schools not following the 11 0.25 .. ..
Swedish curriculum
Survey Population 4 261 98.2 .. ..
Chinese Taipei Target Population 2 656 100 93 713 100
Exclusions: Schools with fewer than three teachers 1 0.04 3 0
Survey Population 2 655 99.96 93 710 100
Turkey Target Population 24 755 100 289 681 100
Survey Population 24 755 100 289 681 100
United Arab Target Population 602 100 22 074 100
Emirates
Exclusions: Private schools 14 2.33 428 1.94
Survey Population 588 97.67 21 646 98.06
Viet Nam Target Population 15 169 100 395 620 100
Exclusions: Non-Vietnamese international schools 26 0.17 685 0.17
Survey Population 15 143 99.83 394 935 99.83
Participating Population and Reasons for exclusions Number of Percentage Number of Percentage
country/economy coverage schools of total Teachers of total
Alberta (Canada) Target Population 688 100 .. ..
Exclusions: Band-operated schools. These 37 5.38 .. ..
schools operate on First Nations’
reserves and are the
responsibility of the federal
government rather than the
responsibility of Alberta
Education
Very small schools (fewer than 45 6.54 .. ..
three students in each of
Grades 10 to 12)
Survey Population 606 88.08 .. ..
Brazil Target Population 28 815 100 643 204 100
Exclusions: Schools with fewer than six 804 2.79 2 682 0.42
teachers. Because ISCED level 2
requires at least one teacher for
each subject, most of these
schools have only one class. The
schools that fit this criterion are
located in geographically remote
areas
Survey Population 28 011 97.21 640 522 99.58
Croatia Target Population 385 100 .. ..
Exclusions: National minority schools (Italian, 6 1.56 .. ..
Serbian, Hungarian)
Survey Population 379 98.44 .. ..
Denmark Target Population 421 100 .. ..
Exclusions: Small schools (fewer than 24 5.7 .. ..
40 students and generally fewer
than 5 teachers)
Survey Population 397 94.3 .. ..
Portugal Target Population 871 100 36 849 100
Exclusions: Schools with non-Portuguese 17 1.95 230 0.62
curricula
Survey Population 854 98.05 36 619 99.38
Slovenia Target Population 153 100 .. ..
Exclusions: Italian upper secondary schools 3 1.96 .. ..
in an ethnically mixed area where
the language of instruction is
Italian and where the teachers
are therefore a separate group of
Slovenian teachers
Survey Population 150 98.04 .. ..
Sweden Target Population 1 296 100 .. ..
Exclusions: Small schools (fewer than 7 0.54 .. ..
six students at ISCED level 3)
International schools not 11 0.85 .. ..
following the Swedish curriculum
Survey Population 1 278 98.61 .. ..
Participating Population and Reasons for exclusions Number of Percentage Number of Percentage
country/economy coverage schools of total Teachers of total
Chinese Taipei Target Population 505 100 .. ..
Exclusions Schools with fewer than 2 0.4 .. ..
three teachers
Survey Population 503 99.6 .. ..
Turkey Target Population 9 520 100 325 692 100
Survey Population 9 520 100 325 692 100
United Arab Target Population 448 100 13 260 100
Emirates
Exclusions: Private schools 11 2.46 285 2.15
Survey Population 437 97.54 12 975 97.85
Viet Nam Target Population 2 941 100 167 755 100
Exclusions: Non-Vietnamese international 13 0.44 156 0.09
schools
Survey Population 2 928 99.56 167 599 99.91
The statistical data for Israel are supplied by and under the responsibility of the relevant Israeli authorities. The
use of such data by the OECD is without prejudice to the status of the Golan Heights, East Jerusalem and Israeli
settlements in the West Bank under the terms of international law.
Table A G.1. Unweighted and weighted participation rates (%) for ISCED level 1 schools
and principals
Table A G.2. Unweighted and weighted participation rates (%) for ISCED level 1 teachers
Unweighted Weighted
Participating
country/ Teachers in Overall Teachers in Overall
Before After Before After
participating teacher participating teacher
economy replacement replacement replacement replacement
schools participation schools participation
Australia 46.2 71.2 76.5 54.5 48.8 74.0 76.4 56.5
Flemish 67.0 88.5 92.0 81.4 66.3 88.5 92.2 81.7
Community
(Belgium)
Ciudad Autónoma 81.0 83.5 86.9 72.5 79.5 81.9 86.2 70.6
de Buenos Aires
(Argentina)
Denmark 58.6 77.8 87.5 68.1 58.4 77.7 87.3 67.8
England 66.3 80.0 85.7 68.6 74.3 85.9 85.0 73.1
(United Kingdom)
France 89.2 91.3 90.7 82.8 88.6 91.2 92.1 84.0
Japan 97.0 99.5 98.8 98.3 97.1 99.5 98.7 98.3
Korea 86.0 91.0 91.9 83.6 85.8 90.3 92.0 83.1
Netherlands 39.2 67.0 86.8 58.2 38.7 67.0 87.3 58.5
Spain 99.3 99.5 95.4 95.0 99.3 99.5 94.1 93.6
Sweden 90.0 93.7 78.8 73.8 89.8 93.8 78.6 73.7
Chinese Taipei 99.5 100.0 97.6 97.6 99.9 100.0 97.0 97.0
Turkey 99.4 99.4 98.5 97.9 99.2 99.2 98.4 97.6
United Arab 99.6 99.6 96.6 96.2 99.7 99.7 96.3 96.0
Emirates
Viet Nam 100.0 100.0 98.0 98.0 100.0 100.0 98.3 98.3
Table A G.3. Unweighted and weighted participation rates (%) for ISCED level 2 schools
and principals
1. Note by Turkey: The information in this document with reference to “Cyprus” relates to the southern part of
the Island. There is no single authority representing both Turkish and Greek Cypriot people on the Island.
Turkey recognises the Turkish Republic of Northern Cyprus (TRNC). Until a lasting and equitable solution is
found within the context of the United Nations, Turkey shall preserve its position concerning the “Cyprus
issue”.
2. Note by all the European Union Member States of the OECD and the European Union: The Republic of
Cyprus is recognised by all members of the United Nations with the exception of Turkey. The information in
this document relates to the area under the effective control of the Government of the Republic of Cyprus.
Table A G.4. Unweighted and weighted participation rates (%) for ISCED level 2 teachers
Unweighted Weighted
Participating Teachers in Overall Teachers in Overall
country/economy Before After Before After
participating teacher participating teacher
replacement replacement replacement replacement
schools participation schools participation
Alberta (Canada) 51.8 62.6 83.0 52.0 48.9 60.3 84.2 50.7
Australia 50.3 76.6 77.7 59.6 48.5 75.1 77.5 58.2
Austria 85.9 88.8 84.4 75.0 85.7 88.6 84.3 74.6
Belgium 85.2 94.2 86.2 81.2 86.0 95.1 86.8 82.6
Flemish 80.0 91.0 84.4 76.8 80.1 91.0 84.2 76.6
Community
Belgium)
Brazil 85.6 94.9 94.6 89.8 89.9 96.6 94.9 91.6
Bulgaria 97.5 100.0 98.1 98.1 97.1 100.0 98.3 98.3
Ciudad Autónoma 81.3 86.7 88.6 76.8 80.7 85.2 89.2 76.0
de Buenos Aires
(Argentina)
Chile 80.5 91.8 94.1 86.3 82.6 91.5 94.3 86.2
Colombia 73.9 77.4 93.4 72.3 73.1 77.1 93.5 72.1
Croatia 94.9 95.9 87.0 83.5 95.4 96.2 87.0 83.7
Cyprus1 88.9 88.9 90.1 80.1 88.9 88.9 90.1 80.1
Czech Republic 100.0 100.0 93.8 93.8 100.0 100.0 93.8 93.8
Denmark 50.5 71.9 86.8 62.5 51.1 72.0 86.8 62.5
England 68.2 77.6 83.1 64.5 72.7 81.5 83.6 68.1
(United Kingdom)
Estonia 88.2 100.0 95.4 95.4 86.6 100.0 95.2 95.2
Finland 100.0 100.0 95.9 95.9 100.0 100.0 96.2 96.2
France 87.9 88.4 87.8 77.6 87.3 87.8 88.1 77.3
Georgia 99.5 99.5 95.8 95.3 99.5 99.5 95.8 95.3
Hungary 94.8 97.9 94.5 92.5 94.9 97.7 95.0 92.8
Iceland 89.7 89.7 75.5 67.7 89.7 89.7 75.5 67.7
Israel 85.3 87.3 84.9 74.2 84.9 86.4 84.9 73.4
Italy 91.7 99.0 93.5 92.5 92.8 99.1 93.8 93.0
Japan 92.4 99.5 99.0 98.5 92.5 99.5 99.0 98.5
Kazakhstan 99.7 100.0 99.7 99.7 100.0 100.0 99.8 99.8
Korea 70.5 81.5 92.2 75.1 69.9 82.5 91.9 75.8
Latvia 79.7 91.2 87.6 79.9 77.1 91.2 87.9 80.2
Lithuania 100.0 100.0 97.4 97.4 100.0 100.0 97.4 97.4
Unweighted Weighted
Participating Teachers in Overall Teachers in Overall
country/economy Before After Before After
participating teacher participating teacher
replacement replacement replacement replacement
schools participation schools participation
Malta 94.8 94.8 86.5 82.0 92.5 92.5 86.5 80.0
Mexico 90.5 96.5 93.9 90.6 90.4 96.3 94.3 90.8
Netherlands 52.7 79.5 80.5 63.9 56.7 79.5 80.9 64.3
New Zealand 63.4 79.3 79.0 62.6 62.6 79.3 79.6 63.2
Norway 77.0 92.5 82.7 76.5 77.4 92.6 83.2 77.0
Portugal 97.5 100.0 91.9 91.9 97.9 100.0 92.7 92.7
Romania 100.0 100.0 98.3 98.3 100.0 100.0 98.3 98.3
Russian 99.1 100.0 99.8 99.8 98.7 100.0 99.9 99.9
Federation
Saudi Arabia 89.9 89.9 85.4 76.8 89.7 89.7 86.0 77.1
Shanghai (China) 100.0 100.0 99.5 99.5 100.0 100.0 99.5 99.5
Singapore 98.2 100.0 99.2 99.2 96.9 100.0 99.1 99.1
Slovak Republic 82.4 88.4 95.0 84.0 82.4 88.9 95.4 84.7
Slovenia 82.7 88.0 91.4 80.5 82.2 88.0 91.5 80.5
South Africa 92.3 92.9 89.7 83.3 92.1 92.4 89.1 82.3
Spain 99.5 100.0 94.6 94.6 99.7 100.0 93.2 93.2
Sweden 90.1 93.8 81.0 76.0 89.1 93.9 81.3 76.3
Chinese Taipei 99.0 99.0 97.2 96.2 98.9 98.9 97.2 96.2
Turkey 99.0 99.0 98.2 97.2 99.0 99.0 98.5 97.5
United Arab 100.0 100.0 96.0 96.0 100.0 100.0 95.7 95.7
Emirates
United States 58.4 77.1 89.1 68.7 60.1 76.8 89.6 68.8
Viet Nam 100.0 100.0 96.3 96.3 100.0 100.0 96.1 96.1
Table A G.5. Unweighted and weighted participation rates (%) for ISCED level 3 schools
and principals
Table A G.6. Unweighted and weighted participation rates (%) for ISCED level 3 teachers
Unweighted Weighted
Participating Teachers in Overall Teachers in Overall
country/economy Before After Before After
participating teacher participating teacher
replacement replacement replacement replacement
schools participation schools participation
Alberta (Canada) 49.7 58.0 79.6 46.2 51.6 56.6 80.2 45.4
Brazil 87.0 96.9 94.2 91.3 92.2 97.4 94.5 92.0
Croatia 98.0 98.0 89.5 87.7 97.9 97.9 89.7 87.9
Denmark 65.5 79.9 85.7 68.5 72.2 85.6 84.7 72.4
Portugal 98.0 99.5 91.2 90.7 99.0 99.7 91.3 91.0
Slovenia 80.4 80.4 87.8 70.6 80.4 80.4 87.7 70.5
Sweden 94.6 97.8 80.8 79.1 95.3 97.8 81.7 79.9
Chinese Taipei 98.0 98.0 95.9 94.0 98.1 98.1 95.8 94.1
Turkey 100.0 100.0 97.7 97.7 100.0 100.0 98.0 98.0
United Arab 99.3 99.3 95.7 95.0 99.0 99.0 95.4 94.5
Emirates
Viet Nam 100.0 100.0 97.5 97.5 100.0 100.0 97.7 97.7
Table A G.7. Unweighted and weighted participation rates (%) for TALIS-PISA link schools
and principals
Table A G.8. Unweighted and weighted participation rates (%) for TALIS-PISA link
teachers
Unweighted Weighted
Participating Teachers in Overall Teachers in Overall
country/economy Before After Before After
participating teacher participating teacher
replacement replacement replacement replacement
schools participation schools participation
Australia 66.9 88.5 92.5 81.9 65.6 88.8 93.4 82.9
Ciudad Autónoma 87.7 90.1 85.1 76.7 86.5 88.8 84.4 74.9
de Buenos Aires
(Argentina)
Colombia 91.4 95.1 94.6 89.9 90.4 93.9 95.1 89.3
Czech Republic 100.0 100.0 93.2 93.2 100.0 100.0 94.8 94.8
Denmark 63.3 66.7 82.4 54.9 65.8 70.0 85.9 60.2
Georgia 91.7 91.7 94.0 86.2 93.1 93.1 94.3 87.8
Malta 88.0 88.0 88.6 78.0 86.7 86.7 88.6 76.8
Turkey 96.6 96.6 99.4 96.0 97.9 97.9 99.6 97.5
Viet Nam 99.1 99.1 98.5 97.7 99.3 99.3 98.4 97.7
Annex H. Questionnaires
Table A I.1. Descriptive statistics of the scaled scores (based on all populations, unweighted)
Note: Participating countries/economies with untrustworthy scale scores were exclude from these averages.
Source: OECD, TALIS 2018 database.
Table A I.2. Country-specific invariance levels across ISCED levels within countries that participated in more than one ISCED level:
Principal scales
Participating
T3PACAD T3PCOM T3PDELI T3PDIVB T3PLACSN T3PLEADP T3PLEADS T3PORGIN T3PWLOAD T3PJOBSA T3PJSENV T3PJSPRO
country/economy
Alberta (Canada)1 Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Flemish Community Scalar Scalar Metric Scalar Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar
(Belgium)1
Brazil2 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Ciudad Autónoma de Scalar Scalar Metric Scalar Metric Metric Scalar Scalar Metric Configural Scalar Configural
Buenos Aires
(Argentina)2
Croatia2 Scalar Metric Metric Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Denmark Scalar - Metric Scalar Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar
England Scalar Scalar Metric Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar
(United Kingdom)1
France1 Metric Metric Configural Scalar Scalar Metric Metric Scalar Metric Metric Metric Scalar
Japan1 Scalar Metric Metric Scalar Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar
Korea1 Scalar Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar Metric Metric Scalar
Portugal2 Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Slovenia2 Scalar Metric Metric Scalar Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar
Spain1 Scalar Metric Metric Scalar Metric Metric Scalar Scalar Scalar Metric Metric Scalar
Sweden Scalar - Metric Scalar Scalar Metric Scalar Scalar - Scalar Scalar* Scalar
Chinese Taipei Scalar Metric Configural Scalar Metric Metric Scalar Scalar Scalar Scalar Scalar Scalar
Turkey Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
United Arab Emirates Scalar Scalar Configural Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Viet Nam Scalar Metric Configural Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
1. Country/economy participated in ISCED levels 1 and 2.
2. Country/economy participated in ISCED levels 2 and 3.
All participating countries/economies that do not have a note participated in ISCED levels 1, 2 and 3.
* The ISCED 3 population for Sweden was not included in the measurement invariance testing for this scale.
Note: A dash (-) indicates that the country/economy was excluded from the scale due to model non-convergence in the evaluation steps.
Source: OECD, TALIS 2018 database.
Table A I.3. Country-specific invariance levels across ISCED levels within countries that participated in more than one ISCED level:
Teacher scales
Participating
T3DISC T3DIVP T3EFFPD T3PDBAR T3PDIV T3PDPED T3PERUT T3SECLS T3SATAT T3SOCUT T3STAKE T3STBEH T3STUD T3COOP T3EXCH T3COLES
countries/economies
Alberta (Canada)1 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Scalar Metric
Flemish Community Metric Scalar Scalar Metric Scalar Scalar Scalar Scalar Metric Scalar Metric Scalar Metric Metric Scalar Metric
(Belgium) 1
Brazil2 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Ciudad Autónoma Scalar Scalar Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
de Buenos Aires
(Argentina)2
Croatia2 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Scalar Metric
Denmark Metric Scalar Scalar Metric Metric Metric Scalar Metric Metric Metric Metric Scalar Metric Metric Metric Metric
England (United Scalar Scalar Scalar Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar Metric Scalar Metric Scalar Metric
Kingdom)1
France1 Metric Scalar Scalar Metric Scalar Scalar Scalar Scalar Metric Scalar Metric Metric Metric Metric Scalar Metric
Japan1 Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Scalar Metric
Korea1 Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Scalar Scalar Metric Scalar Metric Metric Metric
Portugal2 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Scalar Scalar Scalar Scalar
Slovenia2 Scalar Scalar Scalar Metric Scalar Metric Metric Scalar Scalar Scalar Metric Metric Scalar Metric Scalar Metric
Spain1 Metric Scalar Scalar Metric Scalar Scalar Scalar Scalar Metric Metric Scalar Metric Scalar Metric Metric Scalar
Sweden Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Metric Scalar Metric Metric Scalar Metric
Chinese Taipei Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Scalar Scalar Scalar Scalar Scalar
Turkey Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Metric Scalar
United Arab Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Emirates
Viet Nam Scalar Scalar Scalar Scalar Metric Scalar Scalar Scalar Metric Scalar Scalar Metric Scalar Metric Scalar Metric
Table A I.4. Country-specific invariance levels across ISCED levels within countries that participated in more than one ISCED level
(teacher scales continued)
Participating
T3TEAM T3VALP T3WELS T3WLOAD T3SELF T3SEENG T3SEFE T3SEINS T3TPRA T3CLAIN T3CLASM T3COGAC T3JOBSA T3JSENV T3JSPRO
countries/economies
Alberta (Canada)1 Scalar Metric Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Flemish Community Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Metric Scalar Metric Scalar Scalar Scalar
(Belgium) 1
Brazil2 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Ciudad Autónoma de Scalar Scalar Scalar Metric Metric Metric Scalar Scalar Metric Metric Metric Metric Scalar Scalar Scalar
Buenos Aires
(Argentina)2
Croatia2 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Denmark Scalar Scalar Scalar Metric Metric Metric Metric Metric Metric Metric Metric Metric Scalar Scalar Scalar
England (United Scalar Scalar Scalar Metric Metric Metric Metric Scalar Metric Scalar Metric Metric Scalar Scalar Scalar
Kingdom)1
France1 Scalar Scalar Scalar Metric Metric Metric Scalar Metric Metric Scalar Metric Metric Scalar Scalar Scalar
Japan1 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Metric Metric Metric Scalar Scalar Scalar
Korea1 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Metric Metric Metric Scalar Scalar Scalar
Portugal2 Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Metric Scalar Metric Scalar Scalar Scalar Scalar
Slovenia2 Scalar Scalar Scalar Metric Scalar Scalar Scalar Scalar Metric Metric Scalar Scalar Scalar Scalar Scalar
Spain1 Scalar Scalar Scalar Scalar Metric Metric Scalar Scalar Metric Scalar Scalar Metric Scalar Scalar Scalar
Sweden Scalar Scalar Scalar Metric Metric Metric Metric Scalar Metric Metric Scalar Metric Scalar Scalar Scalar
Chinese Taipei Scalar Scalar Scalar Metric Scalar Scalar Scalar Scalar Metric Scalar Scalar Metric Scalar Scalar Scalar
Turkey Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
United Arab Emirates Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar Scalar
Viet Nam Scalar Scalar Scalar Scalar Metric Scalar Metric Scalar Metric Scalar Scalar Metric Scalar Scalar Scalar
Recoded variables that are part of the TALIS 2018 public data set
1 = Under 25
2 = 25-29
3 = 30-39
4 = 40-49
5 = 50-59
6 = 60 and above
1 = Under 40
2 = 40-49
3 = 50-59
4 = 60 and above
Recoded variables that are not part of the TALIS 2018 public data set
Note
1
ISCED code 1997 level 5 was split into level 5 (short tertiary), 6 (tertiary, bachelor’s), 7 (tertiary,
master’s) in ISCED code 2011. Therefore, ISCED code 1997 level 6 became ISCED code 2011
level 8 (doctoral or equivalent): http://uis.unesco.org/sites/default/files/documents/international-
standard-classification-of-education-isced-2011-en.pdf (UNESCO-UIS, 2012[1]).
References
The Teaching and Learning International Survey (TALIS) asks teachers and school leaders about working
conditions and learning environments at their schools to help countries face diverse challenges. This technical
report details the steps, procedures, methodologies, standards and rules that the TALIS 2018 cycle used
to collect high-quality data. The primary purpose of the report is to support readers of the international and
subsequent thematic reports, as well as users of the public international database when interpreting results,
contextualising information and using the data.