Skip to main content

A comparability study between the General English Proficiency Test- Advanced and the Internet-Based Test of English as a Foreign Language



This study examined the comparability of reading and writing tasks of two English language proficiency tests—the General English Proficiency Test-A (GEPT-A) developed by Language Training Center, Taipei and the Internet-Based Test of English as a Foreign Language (iBT) developed by Educational Testing Service, Princeton.


Data was collected from 184 test takers, 92 in Taiwan and 92 in the USA. Three specific analyses were conducted: First, a content analysis was performed on the passages in the GEPT-A and on the iBT reading passages. Second, a task analysis of the construct coverage, scope, and task formats used in the reading sections, and third, a test performance analysis of scores on the two tests were conducted.


The results of the text analysis showed the reading passages on the two tests are comparable in many ways but differ in several key regards. The task analysis revealed that the construct coverage, item scope, and task formats of the two tests are clearly distinct. Analysis of test performance showed that scores on the GEPT-A and iBT are highly inter-correlated with each other. Exploratory and confirmatory factor analyses of the test score data indicated that the two tests appeared to be measuring reading and writing ability but emphasize different aspects of the reading construct.


Although the two tests are comparable in many ways, the reading passages differ in several key regards. Analyses of participant responses indicated that the two tests assess the same reading construct but emphasize different aspects of it.


This study examined the comparability of reading tasks in two language tests, the General English Proficiency Test-Advanced (GEPT-A) and the Internet-Based Test of English as a Foreign Language (TOEFL) (iBT). A study of this kind is valuable for many reasons. First, although the two tests have similar purposes (that is, the tests are used for selection and placement into undergraduate and graduate English-medium universities in the USA and Canada), the two customer bases are different (the GEPT-A is primarily a test taken in Taiwan whereas iBT is taken worldwide). This difference could be reflected in the tests. Second, a study comparing the content and tasks as well as test performance on the two tests should be required research before test agencies develop concordance score tables and use the tables for making decisions and portability of tests and scores. Comparability in terms of tasks and construct coverage and factor structures based on test performance are essential conditions before both test scores can be used in a concordance table. Without this evidence to support the comparability of the tests, concordance tables would be meaningless and misleading. Third, a study of this kind could also be used to evaluate the claims of the tests themselves: analyses of the test content and test performance are essential aspects of such evaluations of assessments (Kunnan, 2017). Specifically, this study examined the depth to which the reading tasks are comparable and the degree to which they measured similar language abilities in reading. This study was undertaken for these reasons.

Previous relevant studies

The Cambridge-TOEFL comparability study (Bachman, Davidson, Ryan and Choi, 1995)

This widely known comparability study of language tests was conducted in 1987–1990 and published in 1995 by Lyle Bachman and colleagues. It investigated the comparability of the First Certificate of English (FCE) administered by the University of Cambridge Local Examinations Syndicate (now Cambridge ESOL) and the paper-and-pencil version of the Test of English as a Foreign Language (TOEFL) administered by the Educational Testing Service, Princeton.

The research steps used were twofold: a qualitative content analysis of the items, tasks, and prompts and a quantitative analysis of the test performance of the subjects on the tests. The most important aspect of the study related to the care taken in choosing the test instruments, test samples (in terms of test taker characteristics and test takers’ score norms), and administrative and scoring procedures in such a way that they truly represented the two testing practices. The instruments used were authentic FCE tests, institutional (retired) TOEFL and SPEAK, the retired form of The Test of Spoken English, and a modified version of the Test of Written English. Sampling procedures included selecting participants that represented the characteristics of the test takers of the two tests. In addition, an examination of the descriptive statistics demonstrated that the means and standard deviations of the study subjects and the test norms of worldwide test taker groups for the two tests were only two points apart and were not practically different. Further, the administration and scoring procedures mirrored the procedures used by the test administrators and raters of the two tests. With these strict measures in place, it was possible to make conclusions regarding comparability based on the test content analyses, the test performance analyses, and the correlational analyses.

The qualitative content analysis of the two tests was conducted by expert judges who used the communicative language ability instrument developed for the study. It was concluded that in general, there were more similarities between the two tests than there were differences. The quantitative statistical analysis of the two tests was conducted by analyzing the test performances of the study participants. The procedures included descriptive statistics, reliability analyses, correlational analyses, and factor structure analyses for each of the individual tests and then across the two tests. The study concluded that as the same higher-order factor structure was supported individually for the two tests and across the two tests, the two tests generally measured similar language abilities.

Other comparability studies

Other related comparability studies include the following studies: Bachman, Davidson and Milanovic (1995) performed content analyses using the newly developed Communicative Language Ability (CLA) and Test Method Facet (TMF) frameworks. The study used the framework to conduct a systematic procedure to analyze test content—the linguistic characteristics, test format characteristics, and the communicative language abilities that were tested in the test items, tasks, and prompts. Another study, Kunnan (1995), from the same dataset focused on the relationship between test taker characteristics and test performance on the two tests, the FCE and the TOEFL. This study conducted exploratory and confirmatory factor analyses and structural equation modeling on the test performance data, and it once again showed that the two tests measured similar language abilities.

Another type of study that compared two versions of the same test was the Choi, Kim, and Boo (2003) study. This study compared a paper-based language test and a computer-based language test—the Test of English Proficiency developed at Seoul National University. The findings based on content analysis using corpus linguistic techniques showed that both versions of the test are comparable. Along similar lines, another study that investigated the comparability of conventional and computerized tests of reading in a second language was Sawaki (2001). This study provided a comprehensive review of literature regarding cognitive ability, ergonomics, education, psychology, and L1 reading. The study did not draw clear conclusions and generalizations about computerized language assessments due to the range of characteristics such as administrative conditions, computer requirements, test completion time, and test takers’ affect. Similarly, yet another study investigated the comparability of direct and semi-direct speaking test versions (O’Loughlin, 1997). In general, the author concluded that the live and tape-based versions of the oral interaction sub-tests cannot be substituted for each other.

A different approach to investigating equivalence using psychometric equivalence between tests was conducted by Geranpayeh (1994). The study investigated the comparability of test takers’ scores who took two tests, the TOEFL and the International English Language Testing Service (IELTS). The study found that high to moderate correlations between the TOEFL and IELTS scores were found. Along the same lines, Bridgeman and Cooper (1998) conducted a study to investigate the comparability of scores from hand-written and word-processed essays. Results indicated that the scores were higher on the hand-written essays than on word-processed essays. Other more recent investigations included Weir and Wu (2006) who investigated the task comparability in semi-directed speaking tests of three forms of the GEPT-Intermediate, Stansfield and Kenyon (1992) who investigated the comparability of the oral proficiency interview and the simulated oral proficiency interview, and Hawkey (2009) who compared historical issues and themes in developing the two tests from Cambridge ESOL, the First Certificate of English and the Certificate in Advanced English.

The two Bachman et al. (1995, 1996) studies mentioned above obviously have direct bearing on this study as these studies investigated the comparability of two different tests. Therefore, this study will follow procedures used in the Bachman studies such as conducting both a test content and test performance analysis.

Research questions

The research questions were:

  1. 1.

    What were the content, tasks, and response formats of the reading tasks of the GEPT-A and the iBT in terms of the test passages, items, and tasks?

  2. 2.

    What was the test performance of the study participants on the reading and writing tasks of the GEPT-A and on the iBT?

  3. 3.

    What was the comparability of the factor structures of the GEPT-A and iBT based on reading and writing scores?



The study participants consisted of 184 test takers; 92 from Taiwan and 92 from the USA. All participants were university students; 38% were undergraduates and 62% graduates; 92% of them were between the ages of 17 and 30; 45% male and 55% female; and their main academic majors were 34% engineering, 20% humanities, and 17% business.Footnote 1 They responded to reading and writing tasks on the GEPT-A in a standard administration. In addition, they submitted scores they obtained on the iBT taken within 2 years prior to the study; this was a requirement to participate in the study.



One form of the GEPT Advanced Reading Comprehension Test (Form AR-1101) and Advanced Level Writing Test was used for this study. The reading test consisted of two sections: Careful Reading with four passages and 20 items, including multiple-choice, short answer, and matching items, and Skimming and Scanning with three passages (with the last passage consisting of three short passages) and 20 items. The items included both matching and fixed-frame multiple choice. The writing test consisted of a task that required test takers to read two essays and write an essay in response.

Internet-Based TOEFL (iBT)

A practice iBT test form from The Official Guide to the TOEFL Test (Educational Testing Service, 2012) was used for the content comparison. Each reading test featured nine passages with a total of 122 items including multiple choice, a limited number of multiple-response multiple-choice items (selected response items in which test takers must select two or more correct options per item), and one categorization item. The writing task used was an independent writing task; it required test takers to write an essay in response to a brief prompt.

Data analysis

Content analyses

Table 1 lists the topics of the reading passages from both tests.

Table 1 Reading and listening passage identification key for GEPT-A and iBT

The range of disciplines/topics was examined for both tests. The reading passages were also analyzed linguistically using the Coh-Metrix Web Tool and the Compleat Web VP function of the VocabProfiler. Coh-Metrix is a “web-based software tool” developed to “analyze texts on multiple characteristics” (Graesser, McNamara, & Kulikowich, 2011, p. 224) and “to measure cohesion and text difficulty at various levels of language, discourse, and conceptual analysis” (Crossley, Dufty, McCarthy, & McNamara, 2007). The Coh-Metrix analyses used in this study were 43 variables deemed most useful from 106 generated by the web tool.

Task analysis

The reading items from both tests were examined to determine the aspects of the construct of reading that they seemed most likely to assess. Passages were also classified based on their topics. Additionally, the scope of each item was rated, using a five-point scale of very narrow for any item for which the key information necessary to answer it correctly was found within a single sentence, narrow if the key information was within the space of several sentences, moderate for any item for which the key information could be found within all or almost all of a single paragraph, broad if the key information was spread across more than a single paragraph, and very broad for any item for which the key information was distributed across more than half the text. The final step in the task analysis was to identify the response format for items.

Analysis of participants’ performance data

Descriptive statistics were computed for GEPT and iBT reading and writing scores, and these means and standard deviations were used to examine the test performance of the study participants.Footnote 2

Correlations and exploratory and confirmatory factor analyses

Correlations were calculated among the reading and writing tasks for both tests. An exploratory factor analysis was conducted the GEPT-A and the iBT reading and writing scores. A confirmatory factor analysis was then conducted taking the results of the exploratory factor analysis as a starting point (Model 1). The steps for confirmatory factor analysis (CFA) followed standard procedures outlined in Kunnan (1998). Several CFA models with variables from both tests were submitted for evaluation. Goodness-of-fit of the models was evaluated using the following indices: χ 2, the NFI, NNFI, CFI, and RMSEA.


Content analysis

The topics used in the tests have been listed in Table 1. It shows that a wide range of topics are used although the GEPT-A is narrower in terms of topics and does not use as many science and engineering as the iBT.

The Coh-Metrix and LexTutor analyses of the reading passages resulted in 43 values for each reading passage. The independent samples Mann-Whitney U test performed on the 43 variables indicated that only five variables were significantly different across the two tests. The two tests varied significantly in the following areas:

  1. 1.

    The number of words per passage,

  2. 2.

    The Measure of Textual Lexical Diversity (MTLD) across all words,

  3. 3.

    The Measure of Diversity of Vocabulary (VOCD) across all words,

  4. 4.

    The mean number of modifiers per noun phrase,

  5. 5.

    The mean sentence syntax similarity across paragraphs, and

  6. 6.

    The percentage of K1 words in each passage.

These statistics are presented in Table 2.

Table 2 Significant differences between the GEPT-A and iBT reading passages

Task analysis

The GEPT-A passages came from a range of subject matter topics but with no content from the life sciences. In contrast, the iBT passages covered the same sorts of topics as the GEPT but with the addition of life sciences topics as well. Notably, the iBT physical sciences passages all dealt with geology.

As can be seen in Table 3, the construct coverage of the two tests is similar in some aspects but dissimilar in others. First, in terms of reading for specific details, both tests are not that far apart (27.5 vs. 34.4%). In terms of paraphrasing and/or summarizing, the GEPT-A assesses this more extensively than does the iBT (15.5 vs. 6.6%). Neither does much to assess the ability to identify the main idea of a passage (the GEPT-A included one item). In contrast, the iBT includes a number of items to assess the ability to read for major points or ideas (eight in all, one for each passage). And, the iBT includes many items assessing vocabulary knowledge or the ability to determine the meaning of unfamiliar vocabulary from context (36 in all). Finally, another major difference between the two tests lies in the areas of top-down reading processes such as inferencing, identifying author purpose (17 for both), and sensitivity to rhetorical organization and cohesion (11 for both). The iBT includes these to a far greater extent than does the GEPT-A. However, the GEPT-A devotes a whole section to skimming and scanning (20 in all); the iBT does not have any items of this type.

Table 3 Summary of construct coverage for the GEPT-A and iBT

Table 4 describes the breakdown of the scope of the reading items on the GEPT-A and iBT. As can be seen, the GEPT-A predominantly uses items with a narrow or very narrow scope (i.e., requiring the processing of several sentences or less). The iBT, on the other hand, focuses more on moderate-scope items (i.e., those requiring the processing of an entire paragraph (or close to it)), with this level of scope proving to be the most common one. Finally, the iBT includes a high proportion of items with broad or very broad scope (i.e., the key information was spread across multiple paragraphs or the entire passage, respectively).

Table 4 Summary of scope of reading items for the GEPT-A and iBT

Table 5 summarizes the response formats of the two tests. While the iBT was entirely dependent upon selected response items, the GEPT-A included a substantial proportion of short answer items, with only about a third of the items using traditional multiple choice. In contrast, the iBT mainly relied upon multiple choice items (92.6%).

Table 5 Summary of response formats of reading items on the GEPT-A and iBT

Analysis of participants’ performance

Table 6 provides the descriptive statistics for GEPT-A and iBT scores. Scores are reported in percentages for comparability. The means for the two tests show that the GEPT-A reading and writing tests were more difficult than the iBT reading and writing tasks.

Table 6 Descriptive statistics for GEPT and iBT scores in percentage scores

Correlational analysis

Table 7 shows the correlation matrix for the two GEPT-A and iBT reading and writing scores. Unsurprisingly, all of the correlations among test scores were highly significant (p ≤ .001). The correlation between the two GEPT-A reading sections (Careful reading and Skimming and Scanning) was very high (0.701). The correlation among the iBT scores was somewhat lower (0.545). Correlations across the GEPT-A and iBT scores were lower overall (range from 0.316 to 0.457).

Table 7 Correlations among GEPT section scores and iBT section scores

Exploratory factor analysis

The results of the exploratory factor analysis (EFA) are summarized in Table 8. A single-factor solution provided relatively high loadings for all variables and was both parsimonious and easy to interpret. In contrast, a correlated two-factor solution (r = .593) yielded a first factor that accounted for most of the GEPT-A reading passages (passages 3 to 7), a second factor on which the iBT reading and writing scores loaded (with reading particularly high), and the GEPT-A writing and the remaining two GEPT-A reading passages cross-loaded on both factors, with roughly equal loadings on each. This solution was not easily interpretable. Therefore, the single-factor model was considered the best solution.

Table 8 EFA results for GEPT-A and iBT reading and writing scores

Confirmatory factor analysis

Several competing models were tested. As explained previously, Model 1, shown in Fig. 1, was based on the results of the one-factor EFA solution with the factor assumed to represent both academic reading and writing. All parameter estimates were significant.

Fig. 1

Model 1(1-Factor Academic Reading-Writing), with parameter estimates

Model 2, shown in Fig. 2, featured all reading scores loading on one factor and both writing scores loading on a second factor with correlated errors. All parameter estimates were significant. As the iBT reading and writing scores were self-reported data with some missing data, thus, error terms for these two variables were correlated. This resulted in a noticeable improvement in model fit.

Fig. 2

Model 2a (2-Factor Reading & Writing with correlated errors), with parameter estimates

Model 3, shown in Fig. 3, featured all reading and writing scores on two factors, the GEPT-A and iBT. All parameter estimates were significant. However, there was marginal difference with these estimates compared to Model 2. But, the big difference was in the factor correlations. In Model 2, the factor correlations between Academic Reading and Academic Writing was 0.85 while the factor correlations between GEPT-A and iBT was only 0.70.

Fig. 3

Model 3 (2-Factor GEPT-A & iBT), with parameter estimates

This difference, along with acceptable model fit indices, helped determine that Model 2 was the most acceptable model with the current data. Table 9 presents the Goodness-of-fit indices for all models.

Table 9 Goodness-of-fit summary for CFA models


Research question 1: content, task, and response format analysis

Research question 1 asked “What is the content, task and response format of the reading and writing tasks of the GEPT-A and the iBT in terms of the reading passages, items, tasks?”


There is a major difference between the two tests. The GEPT-A tasks included passages with a range of topical content but had only one passage dealing with the sciences (and none taken from the life sciences). In contrast, the iBT had a much heavier emphasis on the sciences. Interestingly, the only physical or Earth science topic covered by either the GEPT-A or iBT was geology. This difference perhaps reflects the customer bases of the two tests: the GEPT-A’s base is primarily in Taiwan and its test takers may have less interest in science-oriented topics; the iBT customer base, on the other hand, is worldwide and its test takers have a wider range of academic majors and their major disciplines include science and engineering.

There are also many differences between the two tests. First, the GEPT-A had seven reading passages, four for careful reading and three for skimming and scanning; in contrast, the iBT had three reading passages, all for careful reading. Second, in terms of words per passage, the GEPT-A passages averaged nearly 50 words more than the iBT reading passages. Third, the reading passages differed in certain aspects involving vocabulary. The GEPT-A had a higher level of lexical diversity in its passages on two separate measures. The proportion of words from the K1 list was also higher for the GEPT-A. In terms of syntax, the iBT had on average significantly more modifiers per noun phrase. Finally, in terms of syntactic similarity across paragraphs—an indicator of cohesion and/or of ease of processing—the iBT measured higher than the GEPT-A. These differences likely contributed in making the GEPT-A a more difficult test than the iBT. This finding was also shown in the percentage scores reported on the two tests.


One salient difference between the two tests was GEPT-A’s inclusion of skimming/scanning section with 20 items (the same number as on the Careful Reading section) with a 20-min time allocation. The iBT did not include any skimming or scanning items. However, as there is an overall time limit, it is possible that the iBT reading tasks have to be completed in an expeditious fashion, leading to skimming and scanning rather than careful reading.

Another difference was that the GEPT-A did not include any items targeting vocabulary knowledge or the ability to infer the meaning of unfamiliar vocabulary from context, while the iBT made heavy use of vocabulary items (30% of all items). But, given the iBT response format used (multiple-choice glosses of the word or phrase in item), test takers who knew the word could possibly answer without even having to read the passage.

Third, both tests included items that required students to read for specific details, and both tests made use of paraphrasing rather than using identical language in an item and the passage. On the other hand, the careful reading section of the GEPT-A made extensive use of specific detail items (11 out of 20 items) but did not use inferencing or reading for the main idea items.

Both tests required paraphrasing and summarizing of material read, but they differed in their emphasis. The GEPT-A required both paraphrasing and summarizing but with a greater emphasis on summarizing. In contrast, the iBT straddled the boundary between the two to some extent and involved a much smaller degree of the information reduction that is required in summarizing. In addition, the GEPT-A used short-answer tasks to address this portion of the reading construct, while the iBT used multiple choice.

Finally, the iBT appears to have abandoned main idea items in favor of major points. However, the GEPT-A only included one main idea item and no items related to major points. Also, the GEPT-A only included one author purpose item compared to many inference items and author purpose items in the iBT.

These differences in general show that while GEPT-A is focused on assessing general reading, iBT is focused on the more advanced assessing academic reading, that is, the tasks reflect more the tasks students at university are likely to be engaged in.


The reading comprehension items on the GEPT-A and iBT differed substantially in terms of scope. The overwhelming majority of iBT items (76%) had narrow or very narrow scope, that is, the necessary information to answer items correctly was contained within several sentences or just one sentence, respectively. In marked contrast, 75% of the GEPT-A reading items on the form analyzed had moderate, broad, or very broad scope, requiring test takers to extract the necessary information from most or all of a paragraph, more than one paragraph, or more than half the passage, respectively. This could be expected to make the GEPT-A items more challenging.

Response formats

The two tests differed markedly in terms of the response formats they employed. The majority of reading items on the GEPT-A were selected response but over a third were short answer items. The selected response items included a substantial portion that was not multiple choice—although about one third of all items were multiple choice or fixed-format multiple choice, the remaining 30% were matching. On the other hand, the iBT items analyzed relied overwhelmingly on traditional multiple choice, with eight multiple-response multiple-choice items and one categorization item. The limited use by the iBT of selected response task and the more even distribution of response formats on the GEPT-A clearly sets it apart and stands likely to reduce any impact from test method effects on the scores. Furthermore, the use of short-answer items on the GEPT-A clearly reflects more authentic academic tasks (Bachman & Palmer, 1996); such items seem likely to engage communicative language ability more thoroughly and could also do a better job of testing actual reading as opposed to testing mere recognition of the correct answers in the options.

Research question 2: test performance

Research question 2 asked “What is the test performance of the study participants on the reading and writing tests of the GEPT-A and on the iBT?”

Descriptive statistics

Judging from the descriptive statistics for scores, the study participants received a higher proportion of items correct on the iBT than on the GEPT-A. Similarly, test takers performed better on the iBT writing section than on the GEPT-A writing task. Study participants performed similarly on the two sections of the GEPT-A reading, the careful reading and skimming and scanning sections, and test takers’ iBT scores were similar for reading and writing.

Research question 3: comparability of the tests

Research question 3 asked “What is the comparability of the reading and writing tests of the GEPT-A and iBT?”


The significant correlations among the GEPT-A and iBT reading and writing scores indicated that the two tests assessed similar constructs. The results of the EFA further indicated that the GEPT-A and iBT reading and writing tasks assessed substantially the same construct, since all observed variables (the seven passage-based GEPT-A testlets, GEPT-A writing score, and iBT reading and writing scores) loaded on the same common factor.

Factor analyses

The hypothesis that the two tests measure the same constructs was also supported by the results of the CFA, although not with the same factor structure as suggested by the EFA. The CFA found the best fit was a two-factor (Academic Reading and Academic Writing) rather than for a single-factor model (Table 9). The two-factor reading and writing model also fits better than one with separate factors for GEPT-A and iBT.


There are several areas where the two tests are not similar. The content analysis of the reading passages in terms of topics showed that the GEPT-A placed much less emphasis on the reading of scientific or technical topics than the iBT. This is one area in which the two tests seem to not be comparable. Further, in terms of construct coverage, the GEPT-A does not give adequate coverage to aspects of careful reading such as reading for details and paraphrasing/summarizing and ignores inferencing and the ability to determine the meaning of unfamiliar vocabulary from context. At the same time, however, the iBT omits all coverage of skimming and scanning and has many items that function (or can function) as assessments of vocabulary knowledge rather than reading ability.

The tests are also not comparable in terms of the scope of their reading comprehension items. The GEPT-A has a more even distribution in scope across its items than does the iBT, with a much lower proportion of narrow-scope items than the iBT. This seems appropriate for a test that purports to assess English at a high level of proficiency, whereas a greater emphasis on items of narrow and very narrow scope would be appropriate on tests targeting lower proficiency levels. The response formats used on the two tests are also not comparable, most notably due to the extensive use of short answer items on the GEPT-A in contrast to the iBT which only uses multiple-choice answer items. In addition, the score distributions of the two tests were not equivalent in this study, suggesting that the GEPT-A may have been more difficult than the iBT.

In summary, while the passage and task analyses revealed important differences between the two tests, the correlational and factor analyses indicate that the GEPT-A and iBT are both assessing similar reading and writing constructs. It is probably most accurate to say that the two tests assess the same constructs but from somewhat different perspectives with somewhat overlapping and somewhat different construct definitions.


There were several limitations to this study; we will list the main ones. First, the study only focused on one form of the GEPT-A and commercially published practice tests of the iBT reading and writing. In order to generalize about the two tests, several forms of the two tests should be examined in terms of test content and test performance. Second, the study sample was a very small percentage of the test-taking populations for both tests. This study sample needs to be larger and more representative of the two populations. Third, the study participants were a convenience sample who were highly motivated to take part in the study. This is likely to be different from the test-taking populations and this may have had some effect on their performance.Footnote 3 Fourth, while the data from GEPT-A was collected during the project period (2015–16), data from iBT was anywhere from 2014. It would have been preferable to have data on both tests collected at the same time of the project. Fifth, this study examined the test content and test performance for all the study participants as one monolithic group, but it is likely that students in universities with different academic majors or disciplines may experience the tests differently and perform differently. Finally, another limitation was that while item-level data from the GEPT-A was collected from the study participants, only section-level from iBT was submitted by the study participants. This unevenness may have some effect on the factor structure solutions that were observed through the EFA and CFA analyses.

Future research in comparability studies will need to find ways to overcome these limitations so that findings can be more generalizable and valuable to the test-taking community as well as the various stakeholders.


This study examined the comparability of the GEPT-A and iBT using data from test takers in both Taiwan and the USA using one form of the GEPT-A reading and writing sections and commercially published forms of the iBT reading and writing. Three research questions were posed regarding the content of the GEPT-A and iBT reading and writing tests, performance on the two tests, and the comparability of the two tests. We conclude that the passages on the two tests are comparable in many ways, but reading passages differ in several key regards. Analyses of participant responses indicated that the two tests assess the same constructs but emphasize different aspects of the reading construct. Overall, though, it is clear that there is more comparability than difference between these two tests in terms of reading and writing.


  1. 1.

    From the test performance analysis, it was found that this sample differed from the typical pool of iBT test takers: the study participants scored 95.3 overall which was equivalent to roughly the 73rd percentile among 2014 iBT test takers worldwide, and the mean reading score of 24.9 (82.9% of the possible scale points) roughly equivalent to the 69th percentile among 2014 iBT test takers (Educational Testing Service, 2015). It is not surprising that this sample had such a high ability level overall, given that the Taiwan-based participants were taking the GEPT-Advanced version for local use and the USA-based participants had already scored well enough on the iBT to be admitted to US universities.

  2. 2.

    The actual essays of the study participants on the GEPT-A writing were not analyzed, only the scores for the writing were analyzed. Also, only the scores of the iBT writing were analyzed.

  3. 3.

    See Footnote 1 above for a more detailed explanation.


  1. Bachman, LF, Davidson, F, Milanovic, M. (1996). The use of test methods in the content analysis and design of EFL proficiency tests. Language Testing, 13, 125–150.

    Article  Google Scholar 

  2. Bachman, LF, Davidson, F, Ryan, K, Choi, I-C (1995). An investigation of the comparability of the two tests of English as a foreign language: the Cambridge-TOEFL comparability study. Cambridge, U.K.: Cambridge University Press.

    Google Scholar 

  3. Bachman, LF, & Palmer, AS (1996). Language testing in practice. Oxford, U.K.: Oxford University Press.

    Google Scholar 

  4. Bridgeman, B, & Cooper, R (1998). Comparability of scores on word-processed and handwritten essays on the graduate management admissions test. Princeton, NJ: Research report 143: Educational Testing Service.

    Google Scholar 

  5. Choi, I-C, Kim, KS, Boo, J. (2003). Comparability of a paper-based language test and a computer-based language test. Language Testing, 20, 295–320.

    Article  Google Scholar 

  6. Crossley, S. A., Dufty, D. F., McCarthy, P. M., & McNamara, D. S. (2007). Toward a new readability: a mixed model approach. In Proceedings of the 29 th annual conference of the Cognitive Science Society (pp. 197–202). Austin, TX: Cognitive Science Society.

  7. Educational Testing Service. (2012). The official guide to the TOEFL test (4th Ed.). New York: McGraw-Hill.

  8. Educational Testing Service. (2015). Test and score data summary for TOEFL iBT Tests: January 2014 – December 2014 test data. New York: Retrieved from

  9. Geranpayeh, A. (1994). Are score comparisons across language proficiency test batteries justified? An IELTS-TOEFL comparability study. Edinb Working Pap Appl Linguist, 5, 50–65.

    Google Scholar 

  10. Graesser, AC, McNamara, DS, Kulikowich, JM. (2011). Coh-Metrix: providing multilevel analyses of text characteristics. Educational Researcher, 40, 223–234.

    Article  Google Scholar 

  11. Hawkey, R (2009). Examining FCE and CAE. Cambridge, UK: Cambridge University Press.

    Google Scholar 

  12. Kunnan, AJ (1995). Test taker characteristics and test performance: a structural modeling study. Cambridge, UK: Cambridge University Press.

    Google Scholar 

  13. Kunnan, AJ. (1998). An introduction to structural equation modeling for language assessment research. Language Testing, 15, 295–332.

    Article  Google Scholar 

  14. Kunnan, AJ (2017). Evaluating language assessments. New York: Routledge.

    Google Scholar 

  15. O’Loughlin, K. (1997). The comparability of direct and semi-direct speaking tests: a case study. Unpublished Ph.D. thesis. Melbourne: University of Melbourne.

  16. Sawaki, Y. (2001). Comparability of conventional and computerized tests of reading in a second language. Lang Learn Technol, 5, 38–59.

    Google Scholar 

  17. Stansfield, C, & Kenyon, D. (1992). Research on the comparability of the oral proficiency interview and the simulated oral proficiency interview. System, 20, 347–364.

    Article  Google Scholar 

  18. Weir, C, & Wu, R. (2006). Establishing test form and individual task comparability: a case study of a semi-direct speaking test. Language Testing, 23, 167–197.

    Article  Google Scholar 

Download references


The authors would also like to thank Dr. Sun-Young Shin at Indiana University and Dr. Becky Huang at the University of Texas, San Antonio and student research assistants Carol Celestine and Elizabeth Lee Tien from Nanyang Technological University, Singapore. We also want to thank the study participants from California State University, Fullerton; Indiana University, Bloomington; and the University of Texas, San Antonio; and the study participants in Taiwan.


The authors acknowledge project funding from the Language Training and Testing Center in Taipei, Taiwan, awarded to the first author.

Author information




Both authors contributed equally to the research and research report. Both authors read and approved the final manuscript.

Corresponding author

Correspondence to Antony John Kunnan.

Ethics declarations

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Kunnan, A.J., Carr, N. A comparability study between the General English Proficiency Test- Advanced and the Internet-Based Test of English as a Foreign Language. Lang Test Asia 7, 17 (2017).

Download citation


  • Internet-based Test
  • Performance Analysis Tests
  • Require Test Takers
  • Passage Reads
  • Readers Construct