Loading...

Testing Lexicogrammar

An Investigation into the Construct Tested in the «Language in Use» Section of the Austrian Matura in English

by Theresa Weiler (Author)
©2022 Monographs 278 Pages
Series: Language Testing and Evaluation, Volume 45

Summary

This book explores the construct of language in use, specifically as operationalised through different item types in the Austrian Matura (school-leaving exam). Empirical research on some of these item types is scarce. The author reports on a mixed-methods study. The theoretical frameworks employed are Purpura’s (2004) model of language ability and Weir’s (2005) socio-cognitive framework. The findings suggest that the tasks under investigation assess grammatical form and meaning at the sub-sentential and sentential level. Different item types were also found to target different elements of lexicogrammatical competence. The study contributes to understanding the nature of language in use and sheds light on the application of the socio-cognitive framework to the validation of language in use tasks.

Table Of Contents

  • Cover
  • Title
  • Copyright
  • About the author
  • About the book
  • This eBook can be cited
  • Abstract
  • Acknowledgements
  • Contents
  • List of abbreviations
  • List of figures
  • List of tables
  • 1. Introduction
  • 1.1 Research context
  • 1.2 Rationale for the study
  • 1.3 Organisation of this book
  • 2. Literature review
  • 2.1 Validity and validation in language testing
  • 2.1.1 Test validation frameworks
  • The argument-based approach
  • The assessment utilization argument
  • The socio-cognitive framework
  • 2.2 Perspectives on (lexico)grammar
  • 2.2.1 From a form-based to a communication-based view
  • 2.2.2 Purpura’s (2004) model of grammar
  • 2.2.3 The interrelationship between lexis and grammar
  • 2.2.4 Grammar and lexis in the CEFR
  • 2.3 Assessing (lexico)grammar
  • 2.3.1 Item types
  • 2.3.2 Gap-fill tests
  • 2.3.3 Multiple-choice gap-fill
  • 2.3.4 Banked gap-fill
  • 2.3.5 Word formation gap-fill
  • 2.3.6 Editing
  • 2.4 Research questions
  • 2.5 Chapter summary
  • 3. Methodology
  • 3.1 Introduction
  • 3.2 Overview of the test design cycle of the Austrian secondary school-leaving examination for foreign languages
  • 3.3 Item types and test tasks under investigation
  • 3.4 Research design
  • 3.4.1 Ethics
  • 3.5 Strand 1: Expert judgements
  • 3.5.1 The use of expert judgements
  • 3.5.2 Pilot study 1: Piloting the language in use construct grid
  • 3.5.3 Main Study 1: Using expert judgements to inform the construct
  • 3.6 Strand 2: Test-takers’ intro- and retrospection
  • 3.6.1 The use of introspective methods
  • 3.6.2 Pilot Study 2: Using think-aloud protocols and retrospective interviews to inform the construct
  • 3.6.3 Reading comprehension
  • 3.6.4 Conclusion and implications for subsequent studies
  • 3.6.5 Pilot Study 3: Piloting the Skills Questionnaire
  • 3.6.6 Main Study 2: Using the Skills Questionnaire to inform the construct
  • 3.7 Chapter summary
  • 4. Results
  • 4.1 Using expert judgements to inform the construct (Main Study 1)
  • 4.1.1 Results informing the construct of the four item types
  • 4.1.2 Inter-judge reliability
  • 4.1.3 Findings from the feedback questionnaire
  • 4.2 Using the Skills Questionnaire to inform the construct (Main Study 2)
  • 4.2.1 Anchor items
  • 4.2.2 Results from test performance data
  • 4.2.3 Results from the Skills Questionnaire
  • 4.2.4 Results from the feedback questionnaire
  • 4.3 Chapter summary
  • 4.3.1 Aspects of lexicogrammar tested in the different item types
  • 4.3.2 Differences and similarities between the item types
  • 4.3.3 Extent of reading comprehension required to answer test items correctly
  • 5. Discussion
  • 5.1 Introduction
  • 5.2 Aspects of lexicogrammar that are tested by the items in the language in use tasks (RQ1)
  • 5.3 Differences and commonalities in the aspects of lexicogrammar assessed by the different item types (RQ2)
  • 5.3.1 Multiple-choice gap-fill
  • 5.3.2 Banked gap-fill
  • 5.3.3 Editing
  • 5.3.4 Word formation gap-fill
  • 5.4 The amount of reading comprehension that is necessary to answer the test items correctly (RQ3)
  • 5.4.1 Multiple-choice gap-fill
  • 5.4.2 Banked gap-fill
  • 5.4.3 Editing
  • 5.4.4 Word formation gap-fill
  • 5.5 Answering the overarching research question
  • 6. Conclusion
  • 6.1 Introduction
  • 6.2 Summary of the main findings
  • 6.3 Contributions of this study
  • 6.3.1 Theoretical contributions
  • 6.3.2 Methodological contributions
  • 6.4 Implications of this study
  • 6.5 Limitations
  • 6.6 Suggestions for further research
  • References
  • Appendices
  • Series Index

←12 | 13→

List of abbreviations

AHS Allgemeinbildende Höhere Schulen/Academic secondary schools

BHS Berufsbildende Höhere Schulen/Colleges for higher vocational education

BGF Banked gap-fill

CEFR Common European Framework of Reference

ED Editing

LiU Language in use

MCG Multiple-choice gap-fill

OGF Open gap-fill

SQ Skills Questionnaire

WF Word formation

←14 | 15→

List of figures

Fig. 2.1 The Socio-cognitive Framework for Validating Tests (adapted from Weir 2005a)

Fig. 2.2 A socio-cognitive framework for validating reading tests: Context and cognitive validity (adapted from Weir, 2005a, p. 44)

Fig. 2.3 Components of grammatical and pragmatic knowledge (adapted from Purpura, 2004, p. 91)

Fig. 3.1 Austrian Matura test development cycle (adapted fromStandardisierte Reifeprüfung, 2013)

Fig. 3.2 The methods-strands matrix: A typology of research designs featuring mixed methods (Teddlie & Tashakkori, 2006, p. 15)

Fig. 3.3 Overview of the research design

Fig. 3.4 Research design of Strand 1

Fig. 3.5 Strand 2 of the research design

Fig. 3.6 Extent of reading comprehension involved in English in use

Fig. 3.7 Response pattern WF (n=11), Skills Questionnaire options 11, 12, 13

Fig. 4.1 Results for the grid section “required context” for each item type

Fig. 4.2 Results of “required context” (grid Section 1) for each task

Fig. 4.3 Results for the general focus (grid Section 2) for the whole dataset

Fig. 4.4 Results for the general focus (grid Section 2) for every item type

Fig. 4.5 Results for the general focus (grid Section 2) for each task

Fig. 4.6 Lexical, morphosyntactic and cohesive foci at item type level

Fig. 4.7 Feedback Q1: Frequencies

Fig. 4.8 Feedback Q2: Frequencies

Fig. 4.9 Feedback Q3: Frequencies

Fig. 4.10 SQ results per SQ option and item type: Lexicogrammatical options (options 1 to 10)

Fig. 4.11 SQ results from test-takers answering correctly: Required context

Fig. 4.12 Results per item type from test-takers answering correctly: Required context

←15 | 16→

Fig. 4.13 Participants’ responses to feedback question Q3 per task/group

Fig. 5.1 The language in use item types on a form-meaning-continuum

Fig. 5.2 Required context: Results from expert judges and test-takers

Fig. 6.1 A socio-cognitive framework for validating language in use tests: Context and cognitive validity (adapted from Weir, 2005a, p. 44)

←16 | 17→

List of tables

Tab. 3.1 Matura language in use item types investigated

Tab. 3.2 Test tasks investigated in the main studies (Austrian Matura)

Tab. 3.3 Test tasks used in the pilot studies (Into Europe)

Tab. 3.4 Details of the research phases

Tab. 3.5 Extract from the pilot version of the glossary

Tab. 3.6 Extract of the pilot version of the construct grid

Tab. 3.7 Benchmark scales

Tab. 3.8 Coding scheme for Pilot Study 1

Tab. 3.9 Pilot Study 1: Summary of the expert judgements

Tab. 3.10 95 % confidence intervals including continuity correction

Tab. 3.11 Description of participants of Pilot Study 2

Tab. 3.12 Order of completed test tasks per participant

Tab. 3.13 Number of think-aloud protocols collected per test task/item type

Tab. 3.14 Intercoder agreement

Tab. 3.15 Descriptive statistics (Pilot Study 2)

Tab. 3.16 Code/items ratio of think-aloud protocols

Tab. 3.17 Pilot Study 3: Participants’ native languages

Tab. 3.18 Main Study 2: Participants’ native languages

Tab. 3.19 Final number of participants for analysis

Tab. 3.20 Codes for analysis of open FB questions

Tab. 4.1 Agreement on full set of tasks: 81 % vs. 56 % cut-off

Tab. 4.2 Experts’ agreement (min. 56 %) at item type level

Tab. 4.3 Judgement of Task09, Q10: Required context

Tab. 4.4 Collapsed item foci 1.1 and 1.2

Tab. 4.5 Judgement of Task09/Q3 regarding the required context (grid Section 1)

Tab. 4.6 Judgement of Task02, grid section “required context”

Tab. 4.7 Experts’ agreement (min. 56 %) at task level

Tab. 4.8 Lexical, morphosyntactic and cohesive foci tested in the tasks

Tab. 4.9 Cross table of grid Section 2 and Sections 3, 4 and 5: Results for all item types

Tab. 4.10 Cross table of grid Sections 2 and Sections 3, 4 and 5: Results per item type

Tab. 4.11 AC1 estimates per item focus

Details

Pages
278
Publication Year
2022
ISBN (PDF)
9783631866528
ISBN (ePUB)
9783631866535
ISBN (Hardcover)
9783631865408
DOI
10.3726/b18996
Language
English
Publication date
2021 (December)
Published
Berlin, Bern, Bruxelles, New York, Oxford, Warszawa, Wien, 2022. 278 pp., 126 fig. b/w, 50 tables.

Biographical notes

Theresa Weiler (Author)

Theresa Weiler holds a PhD in applied linguistics with a specialisation in language testing from Lancaster University (UK). She works as a language testing professional at the Austrian Federal Ministry of Education, Science and Research, where she is involved in the development of the standardised matriculation examination for modern languages.

Previous

Title: Testing Lexicogrammar