Testing Lexicogrammar
An Investigation into the Construct Tested in the «Language in Use» Section of the Austrian Matura in English
Summary
Excerpt
Table Of Contents
- Cover
- Title
- Copyright
- About the author
- About the book
- This eBook can be cited
- Abstract
- Acknowledgements
- Contents
- List of abbreviations
- List of figures
- List of tables
- 1. Introduction
- 1.1 Research context
- 1.2 Rationale for the study
- 1.3 Organisation of this book
- 2. Literature review
- 2.1 Validity and validation in language testing
- 2.1.1 Test validation frameworks
- The argument-based approach
- The assessment utilization argument
- The socio-cognitive framework
- 2.2 Perspectives on (lexico)grammar
- 2.2.1 From a form-based to a communication-based view
- 2.2.2 Purpura’s (2004) model of grammar
- 2.2.3 The interrelationship between lexis and grammar
- 2.2.4 Grammar and lexis in the CEFR
- 2.3 Assessing (lexico)grammar
- 2.3.1 Item types
- 2.3.2 Gap-fill tests
- 2.3.3 Multiple-choice gap-fill
- 2.3.4 Banked gap-fill
- 2.3.5 Word formation gap-fill
- 2.3.6 Editing
- 2.4 Research questions
- 2.5 Chapter summary
- 3. Methodology
- 3.1 Introduction
- 3.2 Overview of the test design cycle of the Austrian secondary school-leaving examination for foreign languages
- 3.3 Item types and test tasks under investigation
- 3.4 Research design
- 3.4.1 Ethics
- 3.5 Strand 1: Expert judgements
- 3.5.1 The use of expert judgements
- 3.5.2 Pilot study 1: Piloting the language in use construct grid
- 3.5.3 Main Study 1: Using expert judgements to inform the construct
- 3.6 Strand 2: Test-takers’ intro- and retrospection
- 3.6.1 The use of introspective methods
- 3.6.2 Pilot Study 2: Using think-aloud protocols and retrospective interviews to inform the construct
- 3.6.3 Reading comprehension
- 3.6.4 Conclusion and implications for subsequent studies
- 3.6.5 Pilot Study 3: Piloting the Skills Questionnaire
- 3.6.6 Main Study 2: Using the Skills Questionnaire to inform the construct
- 3.7 Chapter summary
- 4. Results
- 4.1 Using expert judgements to inform the construct (Main Study 1)
- 4.1.1 Results informing the construct of the four item types
- 4.1.2 Inter-judge reliability
- 4.1.3 Findings from the feedback questionnaire
- 4.2 Using the Skills Questionnaire to inform the construct (Main Study 2)
- 4.2.1 Anchor items
- 4.2.2 Results from test performance data
- 4.2.3 Results from the Skills Questionnaire
- 4.2.4 Results from the feedback questionnaire
- 4.3 Chapter summary
- 4.3.1 Aspects of lexicogrammar tested in the different item types
- 4.3.2 Differences and similarities between the item types
- 4.3.3 Extent of reading comprehension required to answer test items correctly
- 5. Discussion
- 5.1 Introduction
- 5.2 Aspects of lexicogrammar that are tested by the items in the language in use tasks (RQ1)
- 5.3 Differences and commonalities in the aspects of lexicogrammar assessed by the different item types (RQ2)
- 5.3.1 Multiple-choice gap-fill
- 5.3.2 Banked gap-fill
- 5.3.3 Editing
- 5.3.4 Word formation gap-fill
- 5.4 The amount of reading comprehension that is necessary to answer the test items correctly (RQ3)
- 5.4.1 Multiple-choice gap-fill
- 5.4.2 Banked gap-fill
- 5.4.3 Editing
- 5.4.4 Word formation gap-fill
- 5.5 Answering the overarching research question
- 6. Conclusion
- 6.1 Introduction
- 6.2 Summary of the main findings
- 6.3 Contributions of this study
- 6.3.1 Theoretical contributions
- 6.3.2 Methodological contributions
- 6.4 Implications of this study
- 6.5 Limitations
- 6.6 Suggestions for further research
- References
- Appendices
- Series Index
List of abbreviations
AHS Allgemeinbildende Höhere Schulen/Academic secondary schools
BHS Berufsbildende Höhere Schulen/Colleges for higher vocational education
BGF Banked gap-fill
CEFR Common European Framework of Reference
ED Editing
LiU Language in use
MCG Multiple-choice gap-fill
OGF Open gap-fill
SQ Skills Questionnaire
WF Word formation
List of figures
Fig. 2.1 The Socio-cognitive Framework for Validating Tests (adapted from Weir 2005a)
Fig. 2.3 Components of grammatical and pragmatic knowledge (adapted from Purpura, 2004, p. 91)
Fig. 3.1 Austrian Matura test development cycle (adapted fromStandardisierte Reifeprüfung, 2013)
Fig. 3.3 Overview of the research design
Fig. 3.4 Research design of Strand 1
Fig. 3.5 Strand 2 of the research design
Fig. 3.6 Extent of reading comprehension involved in English in use
Fig. 3.7 Response pattern WF (n=11), Skills Questionnaire options 11, 12, 13
Fig. 4.1 Results for the grid section “required context” for each item type
Fig. 4.2 Results of “required context” (grid Section 1) for each task
Fig. 4.3 Results for the general focus (grid Section 2) for the whole dataset
Fig. 4.4 Results for the general focus (grid Section 2) for every item type
Fig. 4.5 Results for the general focus (grid Section 2) for each task
Fig. 4.6 Lexical, morphosyntactic and cohesive foci at item type level
Fig. 4.7 Feedback Q1: Frequencies
Fig. 4.8 Feedback Q2: Frequencies
Fig. 4.9 Feedback Q3: Frequencies
Fig. 4.10 SQ results per SQ option and item type: Lexicogrammatical options (options 1 to 10)
Fig. 4.11 SQ results from test-takers answering correctly: Required context
Fig. 4.12 Results per item type from test-takers answering correctly: Required context
←15 | 16→Fig. 4.13 Participants’ responses to feedback question Q3 per task/group
Fig. 5.1 The language in use item types on a form-meaning-continuum
Fig. 5.2 Required context: Results from expert judges and test-takers
List of tables
Tab. 3.1 Matura language in use item types investigated
Tab. 3.2 Test tasks investigated in the main studies (Austrian Matura)
Tab. 3.3 Test tasks used in the pilot studies (Into Europe)
Tab. 3.4 Details of the research phases
Tab. 3.5 Extract from the pilot version of the glossary
Tab. 3.6 Extract of the pilot version of the construct grid
Tab. 3.8 Coding scheme for Pilot Study 1
Tab. 3.9 Pilot Study 1: Summary of the expert judgements
Tab. 3.10 95 % confidence intervals including continuity correction
Tab. 3.11 Description of participants of Pilot Study 2
Tab. 3.12 Order of completed test tasks per participant
Tab. 3.13 Number of think-aloud protocols collected per test task/item type
Tab. 3.14 Intercoder agreement
Tab. 3.15 Descriptive statistics (Pilot Study 2)
Tab. 3.16 Code/items ratio of think-aloud protocols
Tab. 3.17 Pilot Study 3: Participants’ native languages
Tab. 3.18 Main Study 2: Participants’ native languages
Tab. 3.19 Final number of participants for analysis
Tab. 3.20 Codes for analysis of open FB questions
Tab. 4.1 Agreement on full set of tasks: 81 % vs. 56 % cut-off
Tab. 4.2 Experts’ agreement (min. 56 %) at item type level
Tab. 4.3 Judgement of Task09, Q10: Required context
Tab. 4.4 Collapsed item foci 1.1 and 1.2
Tab. 4.5 Judgement of Task09/Q3 regarding the required context (grid Section 1)
Tab. 4.6 Judgement of Task02, grid section “required context”
Tab. 4.7 Experts’ agreement (min. 56 %) at task level
Tab. 4.8 Lexical, morphosyntactic and cohesive foci tested in the tasks
Tab. 4.9 Cross table of grid Section 2 and Sections 3, 4 and 5: Results for all item types
Tab. 4.10 Cross table of grid Sections 2 and Sections 3, 4 and 5: Results per item type
Details
- Pages
- 278
- Publication Year
- 2022
- ISBN (PDF)
- 9783631866528
- ISBN (ePUB)
- 9783631866535
- ISBN (Hardcover)
- 9783631865408
- DOI
- 10.3726/b18996
- Language
- English
- Publication date
- 2021 (December)
- Published
- Berlin, Bern, Bruxelles, New York, Oxford, Warszawa, Wien, 2022. 278 pp., 126 fig. b/w, 50 tables.