Pearson Global Scale of English

Page 1

INTRODUCING THE GLOBAL SCALE OF ENGLISH Executive summary The Global Scale of English (GSE) offers a numeric transformation of the levels defined in the Common European Framework of Reference for Languages (CEFR, Council of Europe 2011). It ranges from 10 to 90 and provides an instrument for more granular and detailed measurements of learners’ level and progress than is possible with the CEFR itself, with its limited number of wide levels. In the spirit of the CEFR itself, GSE values may be attributed to component or enabling skills (e.g. listening, spoken interaction, vocabulary range, phonological control) as well as being combined to indicate an overall level. The intention is to encourage the profiling of learners, rather than merely classifying them by level. The GSE is intended to serve: a) as the metric underlying all Pearson English learning, teaching and assessment products b) as a benchmark for English achievement which can be related to other tests, to national examinations, and to instructional materials c) as a benchmark of the level of English needed for a range of professional and academic purposes.

Origins: GSE and PTE Academic The Global Scale of English has its origins in the reporting scale developed for PTE Academic, which is a high-stakes academic English test for students wishing to study at institutions where English is the language of instruction. The scale was designed to report test takers’ results as a numeric integer from 10 to 90 as well as being located in a CEFR level. In the process of aligning PTE Academic scores to the CEFR, the reporting scale became in effect a generic linear transformation of the set of CEFR levels. The alignment was accomplished in four phases, following the procedure suggested by the Council of Europe for building a case to relate test scores to CEFR levels (Council of Europe, 2009; Martyniuk, 2011): 1.

Familiarisation. All individuals involved in the development of the test received intensive prior training in understanding and applying the CEF.

2.

Specification. The selection and definition of language tasks to be included was closely linked to the CEFR, using a checklist to determine for each criterion mentioned in the CEFR publication (a) whether it was applicable and (b) whether it had been applied.

3.

Standardization. Raters and reviewers underwent rigorous training and had to pass exams evaluating their consistency and their agreement with pre-established standards.

4.

Empirical validation. This has two aspects: (a) the test itself should meet requirements for reliability and validity to serve as a measurement instrument and (b) the relation with the CEFR must be supported by statistical data. Reliability was assessed by a number of means, e.g., by computing test takers’ ability estimates from separate calibrations of all odd and all even items: the high correlation (0.90) found between the scores based on these test halves suggested a reliability of 0.95 for the yet unselected field tests. Likewise validity was determined by several means, including comparing scores obtained by native and non-native speakers of various different age groups. Finally, the statistical link between test scores and CEFR levels was obtained by correlating test taker-centred and item-centred approaches. Both test taker responses and items were assessed by teams of expert raters, and the resulting ratings were Rasch analysed1.

1

For further information see research papers on: http://pearsonpte.com/research/Pages/ResearchSummaries.aspx


Turn static files into dynamic content formats.

Create a flipbook
Issuu converts static files into: digital portfolios, online yearbooks, online catalogs, digital photo albums and more. Sign up and create your flipbook.