Skip to main content

Skill profiles of Japanese English learners and reasons for uneven patterns

Abstract

Background

Variations in the skill profiles of learners have become an important research area in recent years. However, there is a lack of empirical research on this topic in Japan. We conducted three studies to address this gap.

Methods

Study 1 investigated the characteristics of the flat and uneven skill profiles of Japanese learners of English using 10 datasets from five standardized four-skill second-language English proficiency tests. Studies 2 and 3 examined the reasons behind learners having these uneven profiles using a convergent mixed-methods approach (Creswell & Plano Clark, Designing and conducting mixed methods research, 2018) that consisted of a questionnaire and an interview, respectively.

Results

The results of Study 1 suggested that a flat profile is uncommon, and that various types of uneven profiles exist across datasets. The most frequently observed uneven profiles were as follows: (a) listening, speaking, and writing are lower than reading (LSW<R); (b) speaking and writing are lower than listening and reading (SW<LR); and (c) speaking is lower than the other three skills (S<LRW). The results of Studies 2 and 3 suggested three key reasons, namely, (a) insufficient practice, (b) particular subskills or processes required to accomplish test tasks, and (c) unfamiliarity with test formats and test-taking environments.

Conclusions

The results demonstrated the importance of considering uneven skill profiles in L2 research and practice. These results have implications for test development, admission, placement, and pedagogy. For example, skill profile information can benefit admission and placement officers when they make a decision and can aid teachers and administrators in planning remedial instructions.

Introduction

Second-language (L2) learners have flat or uneven profiles in terms of their listening, reading, speaking, and writing skills. The Council of Europe (2001, 2009, 2020) acknowledged different types of uneven skill profiles. Learners with a flat (i.e., even) profile have acquired the four skills to an equal level, whereas those with uneven (i.e., jagged or nonflat) profiles have one or more skills that are substantially higher or lower than the others. For example, a learner may have very good reading skills, average listening and writing skills, and poor speaking skills. The use of proficiency levels, such as those in the Common European Framework of Reference for Languages (CEFR), enables researchers and practitioners to compare profiles across tests using the same standard; however, test users must be cognizant that each test differs in terms of test constructs and purposes, and that the test results may not be strictly comparable (Deygers et al., 2018; Harsch, 2018).

According to North (2021), uneven profiles have long been noted and are more commonly observed than flat profiles, even among advanced language users. In the introduction to his workshop that focused on uneven profiles, North (2021) stated that communicative tests are designed to measure skills separately because testing institutions have recognized profile variability. Moreover, a variety of CEFR scales are provided “to facilitate creating a differentiated needs profile―at an appropriate degree of detail” (North, 2021, 0:13:10). Additionally, he observed that little research has been conducted on the percentages of L2 learners with flat and uneven profiles. Uneven profiles can be divided into various types based on their skill levels, such as those with high skills in reading and writing and low skills in listening and speaking, and vice versa. These differences in skill levels may be the product of a variety of reasons, and the exploration of such reasons could provide insights into the factors affecting learners’ varied profiles, such as educational backgrounds, and affective and cognitive processes. The current study focuses on the breakdown of the flat and uneven skill profiles of Japanese learners studying English as a foreign language (EFL) and the reasons for the uneven profiles (see Ma & Winke, 2022, and Pang & Skehan, 2021, for subskill profiles within each skill).

Literature review

Skill profiles can be regarded as one of the important research areas (Harsch, 2014; Hulstijn, 2015; Hulstijn et al., 2012). Skill profiles, or concise descriptions of each learner’s L2 skills, can provide valuable information on learners’ skill balance and imbalance, as well as learning, teaching, program evaluation, and policy making (Choi, 2017). For example, feedback based on skill profiles can inform learners of their strengths and weaknesses and prompt them to modify their learning routines. Skill profiles can be found in the score report of standardized four-skill tests, such as the Test of English as a Foreign Language® (TOEFL) Internet-based test (iBT). Score reports typically include total scores and skill scores in skill-based sections, the latter of which can be translated into skill profiles.

Since the differences in skill levels are expected, the four skills are usually reported separately and are typically only moderately related to each other (Liao et al., 2010; Sawaki & Sinharay, 2018). Such a moderate degree of relationship among skills mirrors the varied and dynamic nature of how they operate together when processing and completing tasks and how they help researchers define L2 proficiency. For example, Powers (2013) reported that listening self-assessment scores were best predicted using the Test of English for International Communication® (TOEIC) four-skill scores. The prediction rate decreased when speaking, writing, or both were removed. Similar findings were obtained for the prediction of the self-assessments of reading, speaking, and writing. These results suggest that the differences across skill scores allow researchers to more comprehensively assess learners’ ability and predict the overall L2 proficiency (Powers, 2013; see also Li & Zhang, 2021).

The recognition of the importance of L2 learners’ variability in the four-skill profiles has led to increased academic interest into this topic. For example, Ginther and Yan (2018) reported based on their cluster analysis that Chinese international undergraduates (2011 and 2012 cohorts) at an American university showed three-skill profiles from the TOEFL iBT, namely, (a) speaking (S) lower than the other three skills (listening [L], reading [R], and writing [W]), expressed as S<LRW; (b) SW<LR (the difference between SW and LR was 21 to 23 points on averageFootnote 1); and (c) all four skills being relatively low equally (balanced low). The university’s admission policy until 2011 was to accept international students with a minimum TOEFL iBT total score of 80. However, the university noticed a problem with the (b) group, which had a lower grade point average (GPA) in their first year than the group that had the other skill profiles. Consequently, an additional admission requirement of obtaining minimum speaking and writing scores of 18 each was implemented in 2013. Thus, applicants with the (b) SW<LR profile were no longer admitted. The results from the 2013 cohort showed three profiles: (a) S<LRW, (c) balanced low, and (d) balanced high, where the four skills were relatively and equally high. The skill profiles were related to first-year academic performance as the (b) SW<LR group had lower GPA means with larger variations than the other groups. The authors speculated that these uneven profiles may have been caused by an intense TOEFL preparation for the reading and listening sections or by participants engaging in cheating.

Other studies also reported some groups of learners having uneven skill profiles (e.g., Bridgeman et al., 2016; Harsch et al., 2017; Vahed, 2021). Bridgeman et al.’s (2016) study reported an extreme TOEFL iBT score difference (about 16 points) between productive and receptive skills (SW<RL) among some Chinese international undergraduates with business majors at an American university. The TOEFL scores were found to better predict students’ GPA when the scores of such learners were not included in the analysis. The authors also inferred that intensive test preparation could yield unbalanced skill profiles.

Harsch et al. (2017) reported that Chinese and Indian groups, who were undergraduate and graduate students in the UK, had slightly higher scores for reading and listening than writing and speaking (14 points at most and an approximate average of 2 points) in the TOEFL iBT. These differences were not observed in the German group.

The review of the abovementioned studies can be summarized as follows. First, it revealed uneven skill profiles with a divergence between productive and receptive skills (SW<LR). Second, such an uneven profile could be a problem because adequate levels of productive abilities are required for success in academic settings (Ginther & Yan, 2018). It could also pose a challenge in predicting learners’ academic performances (Bridgeman et al., 2016). Third, uneven skill profiles were observed among particular groups of learners (Harsch et al., 2017). Fourth, the admission policy of minimum skill score requirements (e.g., of at least 18 each in the TOEFL iBT speaking and writing sections) can affect students’ post-entry academic performance (Ginther & Yan, 2018).

While previous studies have provided insights into the flat and uneven profiles of the four skills among L2 learners of English, and the potential negative effects of uneven profiles on academic performance, its effects among Japanese EFL learners of English warrant additional examination. Japan’s Ministry of Education, Culture, Sports, Science and Technology (MEXT, 2018) observed that at the group level, some Japanese learners had an uneven profile of SW<LR; however, few empirical studies have extensively examined this topic at the individual level. In the MEXT’s study, L2 four-skill English tests were conducted in 2017 among lower and upper secondary school students to examine how the four English skills can be fostered in a balanced way. MEXT administered listening, reading, and writing tests to approximately 60,000 junior and 60,000 senior high school students and speaking tests to 20,000 junior and 10,000 senior high school students (from 600 national and 300 public schools that were randomly selected from the population). Although the results did not focus on skill profiles, they suggested that some students had an uneven profile. For example, approximately 30% of senior high school students obtained a CEFR level of A2 in listening (30.2%) and reading (29.4%), whereas less than 20% obtained the same level in speaking (11.7%) and writing (19.3%). The small percentage of students with B1 or more in all the skills (i.e., 0.4% to 4.1%) suggests that some students had an uneven profile of SW<LR, which may lead to future problems with academic performance (Ginther & Yan, 2018). However, a more detailed examination was needed as the MEXT (2018) analyzed the skill profiles at the group level, and only individual-level analyses could reveal a detailed overall trend of the learners’ profiles. Furthermore, previous studies (e.g., Ginther & Yan, 2018) including by MEXT did not examine the reason behind the uneven profiles among learners with uneven profiles. Therefore, studies examining the underlying factors behind skill profiles are needed.

Research questions and overall study design

We examined Japanese learners’ skill profiles of all possible skill combinations and learners’ self-perceived reasons for the uneven profiles to investigate the general trend of the varied four-skill profiles of Japanese EFL learners of English at the individual level and the underlying reasons for uneven profiles. Since the breakdown of skill profiles would differ across tests and contexts, we systematically examined all possible combinations of such breakdowns using multiple datasets (from five tests and nine different groups of participants) by posing the following two research questions (RQs).

  • RQ1: What are the characteristics of the four-skill profiles of Japanese EFL learners that are frequently observed across datasets?

  • RQ2: What are the reasons perceived by learners for their uneven profiles?

We conducted three studies, as visually presented in Fig. 1. RQ1 was examined in Study 1 (quantitatively) by comparing skill profiles using 10 datasets from five standardized four-skill English proficiency test scores. The datasets were examined to identify the overall patterns in the flat and uneven skill profiles as obtaining common profiles across datasets and tests could indicate the typical characteristics of the target learners. We made our utmost efforts into obtaining as many datasets as possible by contacting agencies and collecting data ourselves, and we analyzed all the datasets in which we could obtain permission for use.

Fig. 1
figure 1

Overall study design based on Creswell and Plano Clark (2018)

RQ2 was investigated in Studies 2 (quantitatively) and 3 (qualitatively) using a questionnaire and an interview. The combined use of quantitative and qualitative methods followed a convergent mixed-methods design (Creswell & Plano Clark, 2018; see Riazi, 2017 for a summary of the different terminology). The convergent mixed-methods design allowed us to utilize the strengths of the quantitative and qualitative approaches. The quantitative-oriented questionnaire in Study 2 enabled us to identify the overall reasons for uneven skill profiles, which were further examined in the qualitative interview in Study 3. Collectively, the insights from the current article would help researchers clarify the importance of skill profiles in L2 research and practice.

Study 1: investigation into skill profiles

Method for Study 1

Study 1 used 10 datasets to identify whether any patterns could be observed in skill profiles across datasets (RQ1). While they were from different tests and participants, they were all test-score data from learners studying English in Japan that were provided by testing institutions. As shown in Table 1, the 10 datasets contained skill scores from five tests: (a) Global Test of English Communication (GTEC) Computer-Based Testing (CBT; https://www.benesse.co.jp/gtec/en/); (b) TOEFL Junior® Comprehensive (TOEFL Junior, hereafter; https://www.ets.org/toefl_junior/content/); (c) the Test of English for Academic Purposes (TEAP; https://www.eiken.or.jp/teap/); (d) TOEFL iBT (https://www.ets.org/toefl/test-takers/ibt/about); and (e) TOEIC (https://www.ets.org/toeic). All tests consisted of the four-skill sections but differed in delivery modes, range of proficiency levels, and purposes (see Table 2). They had a conversion table that related each test skill score to CEFR.

Table 1 Ten datasets in Study 1
Table 2 Five tests used in Study 1

Datasets

Ten datasets of four-skill scores were analyzed separately. Participants in each dataset were a convenience sample of those who took any one or two of those tests and agreed to participate in the data collection. A majority of the participants (95% or more) were Japanese English learners who studied English primarily in Japan. We used all the data in each dataset, and the number of participants in each dataset varied as a result.

Dataset 1

The dataset contained scores from GTEC CBT from senior high school (public and private) students in Japan. A total of 1805 students took the test in March 2018.

Datasets 2 and 3

The datasets contained scores from TOEFL Junior from 2799 Japanese senior high school (public and private) students and 234 Japanese university (public and private) undergraduates. The data were collected between September 2015 and January 2016.

Datasets 4 and 5

The datasets contained scores from TEAP and TOEFL iBT from 100 Japanese private university undergraduates, who completed both tests in December 2013. The datasets differ from the others in that the same group of students took both tests. This allowed us to rigorously examine the skill profiles of the same participants across the tests.

Dataset 6

The dataset contained scores from TOEIC. In total, 106 Japanese undergraduates and graduates from national universities completed the TOEIC L&R and TOEIC S&W between July and May 2009.

Datasets 7 to 10

The datasets contained scores from TOEFL iBT. The data were collected from 2014 to 2017, between August and November (N = 521). Each year, 126 to 138 undergraduate students at a Japanese private medical university participated in this study. The datasets were unique in that they were all derived from students at one university, which enabled us to examine skill profiles from datasets collected at the same school across 4 years.

Analysis

We identified the CEFR levels for each skill to construct learners’ skill profiles and computed their percentage in two steps. First, four-skill scores were converted into the corresponding CEFR levels using a conversion table provided by each test developer. For example, to convert the TOEFL iBT test scores, we used Table 3 that was adapted from Papageorgiou et al.’s (2015) comparison table of TOEFL iBT and CEFR (see Table 1 of their paper); when a learner received 16 points for the listening section in the TOEFL iBT, the corresponding level in the CEFR level was judged to be at the B1 level for listening. This conversion process was repeated for each learner, skill, and dataset.

Table 3 Comparison table between TOEFL iBT scores and CEFR levels in Study 1

Second, the CEFR levels for each skill were summarized manually. For example, if all four skills belonged to the same CEFR level, the skill profile was considered flat or LRSW. All other profiles were considered to be uneven. If R was one level higher than the other skills, the profile was labeled as LSW<R. If R was two levels higher than the other skills, the profile was labeled as LSW<<R (e.g., R is at the B2 level; the other skills are at the A2 level). After all the profiles were summarized into one table, they were classified to examine whether any patterns could be observed among them.

The CEFR levels for each skill, rather than the test scores, were used to compare the analyses across datasets, using the percentages of skill profiles in each dataset due to the unsuitability of the other two other alternatives. The first alternative method is to use percentile scores or standard scores (e.g., z-scores) for each skill. However, this was difficult to apply here because not all tests reported information on percentile scores, and the target test-takers were not the same across the tests. Therefore, standard scores were not comparable across tests.

We also did not use cluster or latent profile analyses or finite mixture modeling (Choi, 2017; Dunn & Iwaniec, 2021; Ginther & Yan, 2018; In’nami & Koizumi, 2021; Roohr et al., 2022). These methods take a person-centered approach and classify learners according to the similarity in their score patterns. These methods contrast with factor analysis, which takes an item-centered approach and classifies items that receive similar responses (see Flaherty & Kiff, 2012; Oberski, 2016). Cluster and latent profile analyses differ in that the former aims to identify the observed groups, whereas the latter aims to identify the latent groups. Finite mixture modeling includes cluster and latent profile analyses and others (Masyn, 2013). Cluster and latent profile analyses were not suitable for this study for two reasons. First, when the standard scores for each skill in a test were analyzed, the skill profiles were likely to be flat at different proficiency levels. For instance, Sawaki and Sinharay (2013) examined learners’ skill profiles for TOEFL iBT data using cluster analysis (e.g., 14,495 test-takers from the April 2007 administration) and found flat profiles at different proficiency levels. Learners who did well on one section were likely to do equally well on another. Furthermore, cluster or latent profile analyses show an overall pattern of the group, with each profile group including various sub-profiles. Since the current study aimed to examine a general pattern of skill profiles of all possible skill combinations, cluster or latent profile analyses were not used. Second, when the difference between skill standard scores was computed, six scores were derived (L-R, L-S, L-W, R-S, R-W, and S-W scores). These scores were interdependent and likely to result in multicollinearity. Therefore, they were not suitable for cluster or latent profile analyses or finite mixture modeling.

In sum, percentile or standard scores (e.g., z-scores) for each skill were not applied in the current study due to the lack of publicly available information on percentile scores or the different groups of examinees taking tests (making the score comparison difficult). Cluster and latent profile analyses or finite mixture modeling was also not applied as flat profiles were more likely to be observed with standard scores. Furthermore, different scores between skills were expected to be highly correlated with each other, a situation that is not suitable for the use of cluster and latent profile analyses or finite mixture modeling. Hence, the use of CEFR conversion tables and percentages was judged to be the most appropriate for this study. Additionally, since our study was descriptive, we did not use statistical significance testing.

Results and discussion for Study 1

Table 4 shows the results of the CEFR levels in each dataset. A wide range of English proficiency of learners was observed, with the highest percentage at the A2 or B1 level for each skill. There were four exceptions to this: In Dataset 2, most learners had an A1 or lower level in speaking (50.88%); in Datasets 6, 8, and 9, most learners had a B2 level in writing (40.57%, 38.89%, and 33.08%, respectively). This indicates variations in the level of mastery across the skills and datasets.

Table 4 Percentages of participants’ CEFR levels for each dataset in Study 1

We found 75 skill profiles across 10 datasets (see Table 13 in Appendix). From the 75 profiles, 15 skill profiles, with 5% or more of the participants having at least one dataset, were selected and are shown in Table 5 in the order of GTEC CBT. For example, learners with higher reading skills than listening, speaking, and writing skills (LSW<R) were found to be 24.42% and 12.65% of all learners in Datasets 1 and 2, respectively.

Table 5 Skill profile types with 5% or more of participants (%) in Study 1

Those with the same CEFR levels across skills were interpreted as having a flat profile (LRSW). Such learners were observed at small percentages, ranging from 1.59 (Dataset 8) to 29.00% (Dataset 4). This indicates that learners with a flat profile are in the minority, corroborating North’s (2021) argument that uneven profiles are more common. Furthermore, compared with other datasets, the percentages were relatively higher for the data with GTEC CBT taken by senior high school students (Dataset 1, 18.94%), with TOEFL Junior taken by senior high school students (Dataset 2, 17.90%), and with TEAP taken by first-year university students (Dataset 4, 29.00%). Since senior high school or first-year university students are the main target group of such tests, their difficulty level is likely to be adequate for most test-takers. If this is the case, the results might indicate that if learners take a test that matches their proficiency level, they are more likely to obtain a flat skill profile than when they take a test that is too challenging or easy for them. This can be tested in future studies.

Among the various uneven profiles, there were four that 10% or more of learners had across datasets: LSW<R (in eight datasets out of 10), SW<LR (in seven datasets), S<LRW (in seven datasets), and flat (in four datasets). The highest percentages of each skill profile were less than 30% (24.42%, 29.00%, 14.29%, and 20.75%, respectively). A limited number of skill profiles (four) across datasets with a small percentage suggests that skill profiles vary across learners and datasets, with no dominant, single profile of skills explaining performance across learners and datasets.

Further analysis across datasets highlighted a lack of consistency in the skill profiles. For example, the data from two different tests from the same participants (Datasets 4 and 5) produced different results. The data from the same tests with different participants also produced different results (within Datasets 2 and 3 and between Datasets 5 and 7 to 10). These results suggest that the skill profiles vary across learners and tests. The results may be explained by each test having different constructs and purposes (Deygers et al., 2018; Harsch, 2018).

Table 6 shows the results according to the lowest skills. For instance, the lowest skill group of LSW includes LSW<R and LSW<<R (not shown in Table 6; see Table 13 in Appendix). The results show that the lowest skill groups with 10% or more of learners across datasets were LSW in nine datasets (e.g., 25.20% in Dataset 1), S in eight datasets, SW in seven datasets, and LRSW (flat) in three datasets. The groups included S in the four-skill profiles and W in the three-skill profiles. The results of having lower productive skills than receptive skills were consistent with MEXT’s 2018 study. The reasons for the lower levels of S and W were explored in Studies 2 and 3.

Table 6 Skill profile types classified according to the lowest skills (%) in Study 1

Table 7 displays the results of the groups with the highest skills. For example, the highest-skill group of R included LSW<R, LS<W<R, and L<S<W<R (the latter two not shown in Table 7; see Table 13 in Appendix). The results indicate that the groups with 10% or more across datasets were R in nine datasets (e.g., 33.42% in Dataset 1), LR in eight datasets, LRW in seven datasets, and LRSW (flat) in three datasets. The groups included R in the four-skill profiles and L in the three-skill profiles. The reasons for the higher R and L are explored below.

Table 7 Skill profile types classified according to the highest skills (%) in Study 1

Some unique features were found for TOEIC in Dataset 6, where the first to third highest-level skills were W (33.96% in Table 7), LRW (20.75%), and LW (11.32%). This was reflected in a higher percentage of LRS as the lowest skill in TOEIC than in the other datasets (16.98% in Table 6). W had the highest scores among the skills, probably because the writing tasks in TOEIC were easier than those in other tests. The TOEIC W section has three task formats, namely, describing a picture (in one sentence, five items, in eight min), writing a reply to an email (length not specified, two items, in 20 min), and writing an argumentative essay (300 words, one item, in 30 min). While the argumentative-essay task could be of appropriate difficulty for university students who are the target audience of the TOEIC test, the former two tasks (picture description and email response) may be relatively easy for them (see https://www.ets.org/toeic/test-takers/speaking-writing/about/content-format/). Such a difference in task difficulty may not be an issue, since the target language used in the domain of TOEIC includes global workplaces, a feature that differentiates TOEIC from other tests. Still, it would be useful to consider this difference when skill profiles are compared across TOEIC and other tests.

We also focused on how many levels of differences commonly existed between skills (e.g., two-level disparities) by classifying all the skill profiles according to the number of level differences. A one-level difference was found for SW<LR and LSW<R. A four-level difference was found in S<<W<L<R. Table 8 indicates that a one-level difference was most frequently observed, ranging from 55.11 (for Dataset 7) to 70.00% (for Dataset 5). This was followed by a two-level difference (7.00% for Dataset 4, to 40.16% for Dataset 7) and a zero-level difference (i.e., a flat profile; 1.60% for Dataset 8, to 29.00% for Dataset 4). A three-level difference (0.00% for Datasets 4 and 5, to 5.99% for Dataset 3) and a four-level difference (0.00% for Datasets 2 to 10, to 0.06% for Dataset 1) were rarely observed.

Table 8 Skill-level differences and percentages of participants in each category (%) in Study 1

One reason for level differences may be the measurement error of tests, as it is caused by differences in test content, suboptimal test-taking environments, learners’ personal issues, and others (e.g., Green, 2020). If the standard error of measurement (SEM) of a test or section is available, it is possible to calculate a range in which learners’ true scores are included at the 95% probability when they repeatedly take the same test. The range can be computed using the score ± (1.96 × SEM). For example, assume that a learner received 16 points in the TOEFL iBT listening section (equivalent to the B1 level in the CEFR, following Table 3), the SEM of the listening section was reported to be 2.38 (Educational Testing Service, 2018, p. 7), and the learner’s true score was estimated to fall between 11 and 21 by 95% chance (16 ± (1.96 × 2.38) = 16 ± 4.66 = 11.34, 20.66; rounding them to the nearest integer, we get 11 and 21). According to Table 3, the learner’s true CEFR level could fall on either the B1 or B2 level. These values suggest that four-skill and CEFR levels may fluctuate between plus or minus one level due to measurement error, which could lead to level differences in skill profiles. Therefore, it could be argued that a one-level difference is not very worrisome because it is within the margin of measurement error. In contrast, two- or three-level differences may indicate skill imbalances and/or other issues related to cognitive and/or affective factors (Goh & Vandergrift, 2021) and may need to be considered carefully.

Study 2: questionnaire study of learners with uneven skill profiles

Study 1 showed that uneven profiles were common, and that they varied in the mastery level of each skill. To examine the reasons for having uneven profiles (RQ2), Study 2 examined the test-takers’ perceptions of profiles via a questionnaire, focusing on a group with clear uneven profiles in TOEFL iBT.

Method for Study 2

The participant groups were the same as those in Datasets 7 to 10 from Study 1. They were the 2014 to 2017 cohorts at a medical university. Study 2 used the data of the skill scores used in Study 1 as well as the questionnaire collected for Study 2.

We first selected students with uneven skill profiles, that is, those with two- or more-level differences across skills (n = 33 to 53; see Table 9), which were based on their skill profiles derived from their TOEFL iBT scores. Hereafter, these students are referred to as the gap group. Students with a one-level difference in skill profiles were not selected because a one-level difference could be due to measurement errors (see Study 1).

Table 9 The gap group’s perceptions toward their skill scores in Study 2

The gap group was further divided into a higher-proficiency gap group (achieving the B2 or higher level judged from their total scores) and a lower-proficiency gap group (achieving the B1 or lower level judged from their total scores) in order to investigate the possible effects of L2 proficiency on perceived reasons for uneven skill profiles. By comparing gap groups of different levels of proficiency, we intended to examine why some learners performed better in one skill than in others.

To examine the reasons behind the uneven profiles (RQ2), we developed a questionnaire (see Additional file 1) which was conducted for 4 years after the participants took the TOEFL iBT and received their score reports each year. The responses were analyzed separately for each year. As our study was descriptive, we did not use statistical significance testing.

Results and discussion for Study 2

Table 9 shows whether learners in the gap group—a selected group of learners with two- or three-level differences in the mastery of skills—thought that there were sections with lower skill scores than other sections (question 1). Overall, more than two-thirds of the learners reported that their scores were uneven across various skills (e.g., 67.92% in 2014 to 84.85% in 2016). As seen in Table 9 and Fig. 2, among such learners, the largest number of learners reported receiving lower scores for S (e.g., 55.56% in 2014). The second-largest number reported receiving lower scores for L (e.g., 50.00% in 2014). When the gap group was subdivided into the higher- and lower-proficiency groups, different patterns were observed. The high-proficiency gap group reported receiving lower scores in S the most and in L, R, and/or W the second most (L, R, and W in 2014; W in 2015; L and W in 2016; and R and W in 2017). The lower-proficiency gap group reported receiving lower scores in L the most (e.g., 57.14% in 2016) and in S the second most (e.g., 42.86% in 2016) or L and S at the same percentage (e.g., 55.17% in 2014). These results suggest that, according to learners’ perceptions, S is their weak skill regardless of the level of their proficiency; L was frequently the weakest skill among lower-proficiency learners relative to reading and writing skills.

Fig. 2
figure 2

Skills that those in the gap group thought received lower scores in TOEFL iBT

Table 10 and Fig. 3 outline the percentages of learners in the gap group who thought they had sections with lower scores (i.e., responded “yes” to question 1) and who answered question 3 (why do you think the scores in these sections were lower? Select all the reasons that apply to your case). The results show that most learners selected options 6 (I was poor at this skill; up to 68.18% in 2015) and 7 (I have not studied the skill much; up to 52.78% in 2014). They indicate that learners with wide skill gaps felt that they received lower scores in certain skills due to their low skills and/or lack of a sufficient amount of time to study. Options 1 (I did not know the iBT test format well; up to 24.32% in 2017) and 2 (I was tense or confused; up to 36.36% in 2015) were also selected consistently.

Table 10 Test-takers’ views toward reasons of having uneven profiles in Study 2
Fig. 3
figure 3

Major reasons why those in the gap group thought received lower scores in TOEFL iBT

As with the overall group, similar response patterns were observed for the lower-proficiency gap group. For example, most learners selected options 6 (I was poor at this skill; up to 72.41% in 2014) and 7 (I have not studied the skill much; up to 51.72% in 2014). Options 1 (I did not know the iBT test format well; up to 26.92% in 2017) and 2 (I was tense or confused; up to 33.33% in 2016) were also selected consistently.

The response patterns were slightly different among the higher-proficiency gap group. Options 1 (I didn’t know the iBT test format well; up to 42.86% in 2014) and 2 (I was tense or confused; up to 71.43% in 2014) were selected by a large percentage of the learners in this group along with options 6 (I was poor at this skill; up to 75.00% in 2016) and 7 (I had not studied the skill much; up to 57.14% in 2014).

These results indicate that learners with uneven skill profiles attribute their lower scores to their low skills and/or lack of a sufficient amount of time to study. This is likely observed in overall-, higher-, and lower-proficiency learners. In other words, most learners with skill gaps see skill scores as a reflection of their ability and/or effort. High-proficiency learners also see them as a reflection of their knowledge about the test format and their own emotional conditions.

Study 3: interview with learners exhibiting uneven skill profiles

Based on quantitative responses to a questionnaire conducted on learners with uneven skill profiles, Study 2 has reported on learners’ perceptions of the reasons for having such profiles. Study 3 further investigated RQ2 by qualitatively analyzing the interview data from learners exhibiting uneven skill profiles.

Method for Study 3

Participants

We selected six participants from the gap group in Study 2 for Dataset 7 (TOEFL iBT in 2014). They exhibited two- or three-level differences in the mastery of skills in the CEFR levels, with wide gaps across skills. Table 11 summarizes the information of the six students. The first four achieved the B1 level in the CEFR as judged by their actual (not self-reported) total scores in the TOEFL iBT. The other two achieved the B2 level. All took the TOEFL iBT between August and November 2014, when they were first-year university students. They were interviewed between July and August 2015, when they were second-year university students, and the gratuity was paid after the interviews.

Table 11 Six participants’ characteristics in Study 3

Interview procedures

The interview was conducted in a semi-structured, face-to-face format in which an interviewer (one of the authors) and an interviewee (i.e., learner) talked in a quiet room. The interviewer obtained informed consent from the learner and recorded their responses using a voice recorder, along with a video when necessary.

The interview consisted of four phases (see Fig. 4). First, learners received a copy of the questionnaire administered in Study 2, with their responses filled in. They were asked to double-check the responses and revise them as needed. Second, they were asked the following questions: (a) How have you studied English from the time you started until your first year at university? (b) Why do you think that you have lower skills, and how have you studied them? (c) What particular aspects of the skills do you think are the most difficult to study? The interviewer asked questions to clarify what the learners meant in their responses in order to elicit more details.

Fig. 4
figure 4

Interview procedures in Study 3

Third, learners took a TOEFL iBT practice test (Educational Testing Service, 2012). Those who had lower scores in the L section in the practice test also took a listening section in the National Center Test for University Admissions (Center Test, hereafter; see National Center for University Entrance Examinations, 2017; Watanabe, 2013) to further highlight their weak areas. The Center Test is a high-stakes university entrance examination used by almost all Japanese universities, including the participants. The Center Test administered in January 2015 was used in the current study as this version was administered nationwide after the current participants entered university. No participants were expected to have taken it prior to data collection. As learners took the tests, they verbally explained their thought processes. The interviewer recorded their behaviors and responses using video and voice recorders. She also observed and took notes to describe how the learners responded to these tests.

Fourth, learners were asked about distinctive test-taking behaviors noticed by the interviewer and were asked to explain why they behaved in such a particular manner. Additionally, when necessary, they were asked to explain how they arrived at the answers to questions that they had answered incorrectly in the test.

Analysis

The interview responses were transcribed verbatim. One of the authors extracted points related to skills and skill profiles from the transcript, summarized them into a table, and described the summary in bulleted sentences. This was repeated while examining all of the data, based on a thematic analysis method (Nowell et al., 2017). Two other authors double-checked the summary to examine if all relevant points were included while excluding irrelevant ones. Two of the authors then described them in a passage.

Results and discussion for Study 3

By analyzing the results of the interviews, we have identified three main factors that explained the learners’ uneven skill profiles with two- or three-level CEFR-level differences across skills: insufficient practice, particular subskills or processes required to accomplish test tasks, and unfamiliarity with test formats and test-taking environments.

Insufficient practice

Speaking (S)

All the interviewees indicated that they spent less time on learning skills on which they scored lower than on the ones on which they scored higher. This tendency was particularly notable for S, which all six participants perceived as (one of) the English skills that yielded exceptionally poor results. Subaru, Takeru, Gaku, and Hina reported that they had insufficient S training in English. Subaru, Yamato, and Hina pointed out that proficiency in S is not usually required to enter a university in Japan. This could have caused students and teachers at senior high schools and cram schools to allot little time to improving their speaking skills. Indeed, Hina said that the EFL classes at her senior high school did not cover S. Arata also indicated that he had few opportunities to discuss daily issues or express opinions in English in EFL classes at his senior high school.

Unlike Hina and Arata, Yamato went to a comprehensive (or combined) junior and senior high school that emphasized improving all four skills related to students’ English ability. Therefore, he had many opportunities to improve his S and other skills. However, as the university entrance examination season neared, English classes focused more on preparing students using past examination papers. As a result, toward the end of his senior high school year, Yamato felt that the opportunities to practice S were severely limited.

The participants’ responses illustrate that the learners and their schools did not see much need or allocate much time to practicing S in English due to the lack of an S section in university entrance examinations. This finding demonstrates that university entrance examinations affect which English skill(s) senior high school students and teachers focus on. In other words, an insufficient amount of time to practice in speaking English was potentially caused by the negative impacts of entrance examinations on teaching and learning (see Green, 2020; Tsushima, 2011). This seemed to be a factor that led to uneven skill profiles.

Listening (L)

Subaru and Gaku perceived their L skills to be the lowest among the four skills, whereas Yamato and Hina perceived them to be the highest. The interview results suggest that this individual difference may have been caused by whether learners spent a long time improving their L skills to pass an advanced level of an English proficiency test, which was more challenging than the Center Test. It should be noted that, except for the Center Test, it is not common for Japanese universities to require applicants to demonstrate their listening skills via tests or certificates.

Both Subaru and Gaku explicitly reported that they practiced L to perform just sufficiently to achieve the level required by the Center Test. As they did not see the need to improve their L beyond that level, they did not allocate the time to do so. Thus, similar to the case of S, Subaru’s and Gaku’s responses indicate that university entrance examinations influence senior high school students’ decisions on which English skill(s) to focus on.

On the other hand, Yamato, Arata, and Hina worked hard on their L to reach a higher level than that required by the Center Test. It seems that they had different learning goals than Subaru and Gaku. For example, Yamato studied to pass the EIKEN Grade Pre-1 (equivalent to the B2 level in the CEFR; https://www.eiken.or.jp/eiken/en/grades/), which is aimed primarily at Japanese university students. Yamato said that preparing for the test enhanced his L. Hina also took the EIKEN test when she was in senior high school. She took grade 2 (equivalent to the B1 level), which is aimed at Japanese senior high school graduates. Finally, Arata’s senior high school required students to take the GTEC for STUDENTS (a paper-based version of GTEC CBT). Although L was not covered in regular EFL classes at his senior high school, he spent time improving his L to pass the test, along with studying R and W.

Reading (R)

Subaru, Takeru, Yamato, and Arata received higher scores in R (relative to other skills) and perceived their R to be the strongest among the four skills. All six participants said that they worked hard to improve their R skills as they were measured during university entrance examinations. Subaru articulated the importance of R to get into a university as follows: “For university entrance examinations, we can confidently say you need the R skill only. Other than that, you sometimes need the W skill.”

The analysis of interview results repeatedly showed that the contents of university entrance examinations affected students’ and schools’ choices regarding which English skill(s) to focus on. This may have led to an imbalance in the development of the four skills. Although the participants took English courses in their first year at the university, they mentioned them little. They intensively studied English for university entrance examinations to attend a medical school, and that experience may have had a stronger impact on them than their first-year courses at the university.

Particular subskills or processes required to accomplish test tasks

To further investigate learners’ imbalanced English skills, they were asked about the skill(s) that were two or three levels lower (S, L, and/or W) than their highest-ranked skill(s). Specifically, they were asked to explain why and in what ways they had difficulty answering questions correctly in sections that required them to use their lower-ranked skill(s) of the TOEFL iBT (see Xi & Norris, 2021, for the test constructs).

Speaking (S)

All six participants had lower S scores than the highest-ranked skill(s) by at least two levels. Takeru and Gaku said that they were not used to speaking extensively in English. This skill is required in the TOEFL iBT S section, wherein 45 s is allotted to independent tasks and 60 s to integrated tasks. Gaku said that he was not used to producing a long monologue in English. Similarly, Takeru said that although he was capable of sustaining a dialogue by speaking one or two sentences during his turn, he could not make a series of sentences and speak for a longer time. Takeru further explained that he was not able to produce an impromptu monologue because he did not know how to organize his logic to construct a speech. Such difficulties may have stemmed from the EFL education he received at secondary and perhaps tertiary levels. According to a survey conducted among junior and senior high school teachers and principals by the Benesse Educational Research and Development Institute (2016), secondary school EFL classrooms in Japan often or sometimes implement pair or group conversations in which students talk in English (79.9% and 46.4% at junior and senior high schools, respectively). Impromptu speeches, in which students talk about themselves or express their opinions, are not as common (42.7% and 29.4%, respectively). At the university where all six participants attended, some instructors had given students opportunities to practice constructing and producing impromptu speeches. However, more could be done to encourage teachers to shift their attention to oral activities.

Gaku, Arata, and Hina mentioned translating Japanese (their first language) into L2 English when speaking English. Hina told the interviewer that she would structure her speech in Japanese first and then translate it into English during situations that provided her with preparation time and required her to speak extensively (i.e., iBT speech or classroom presentation; she would not use a translation method in L, R, or W). She would use this S-translation strategy because she had not studied S enough. Gaku reported that he could not even transform his thoughts easily into Japanese before trying to speak in English. He seemed to be aware that directly translating Japanese into English would not create sensible outcomes. He seemed to think that he should reorganize his thoughts in plainer Japanese to facilitate the translation process and produce a more accurate English outcome. Arata also mentioned the translation, saying that even if he could speak about the topic in Japanese, it was difficult to translate his ideas into English.

One possible reason why some learners used Japanese when developing English speech is the way that English is taught in secondary schools in Japan. According to the 2016 Benesse Educational Research and Development Institute survey, 68.3 to 68.8% of Japanese secondary school teachers of English either often or sometimes use translation (English to Japanese) exercises in classes (pp. 4–5). Most of them (89.4 to 96.1%) often or sometimes provide grammar explanations. As learners have become familiar with the translation method but have had little experience planning and delivering impromptu speeches, it may be natural for them to use Japanese in the process of producing English sentences. Arata also stated that he was not good at speaking even in Japanese in that he could not speak simply and tended to structure his thoughts in a complex manner with much time spent when possible. This suggests that lower S scores can be attributed not only to L2 skills but also to L1 (i.e., Japanese) proficiency.

Listening (L)

Subaru, Takeru, and Gaku stated that when they did not understand part of what was said, they tried to derive the missing information and overall meaning of the speech from the partial information they were able to capture and their background knowledge. However, such an approach was not always successful for long passages or conversations, such as in the TOEFL iBT L section. Subaru said that he could catch parts of words, but he could not always put them together to understand the whole text. Takeru said that sometimes he could understand a part of a speech and would try to guess the whole story based on what he could. However, as he revealed, such an L strategy often resulted in an overly imagined story. He also mentioned that his concentration usually lasted briefly, as he would start to think of something else when he lost interest in what he was listening to.

Another common point among the three learners was that as senior high school students, they practiced L to pass the Center Test but stopped practicing after they reached the passing level. As speeches used for the Center Test are generally shorter than those used in TOEFL iBT, it is not surprising that the learners found the latter’s listening section challenging.

Writing (W)

Takeru and Yamato seemed to have their own explanation for the difficulties presented by W in English. Yamato said that he attempted to write perfectly from the beginning, by writing a complete introduction, and then completing the body and the conclusion. It is not difficult to imagine that if he flexibly started from the body or stopped writing the introduction, and skipped to a different section when he could not come up with “perfect” sentences for the introduction, he could have completed his essays more strategically. He mentioned that similar inflexibility could be observed in his daily life, and that his friends tended to say that he lacked flexibility. Based on his responses, the reason why Yamato struggles with W may be connected to his approach when he writes in English, as well as when he performs tasks both related and unrelated to language.

Takeru could not write a summary of the passage well. He revealed that because of his poor listening skills, he rarely used a listening text, primarily summarizing a reading text in an R-L-and-then-W task. Furthermore, he stated that he could not perform the same R-L-and-then-W task in Japanese. Therefore, Takeru’s lower W score may be related not only to his English skills but also to his W and L skills in Japanese.

Unfamiliarity with test formats and test-taking environments

Some participants were not familiar with the formats for the S, L, and/or W sections, which made it difficult for them to correctly respond to questions.

Speaking (S)

All participants except for Subaru found that the integrated tasks were difficult, partly because they were not familiar with the format. They seemed to be aware of such a format, but they did not practice enough to be comfortable with it. Takeru said he practiced the L-and-then-S type of integrated tasks but did not practice the R-L-and-then-S task before he took the actual TOEFL iBT. Hina said that she took the test without fully understanding the format of the S section. Regarding the integrated tasks, three students (Takeru, Gaku, and Arata) said that they had trouble speaking because they did not understand the L and/or R inputs that they were required to orally summarize and/or give their opinions.

Yamato and Hina revealed that they were not used to accomplishing tasks, such as preparing a structured speech in a short time (i.e., 15 or 30 s). They made similar comments when explaining the difficulties they experienced while answering questions in the S section.

The physical test-taking environments in the test room and technical issues may have also affected the participants’ S performance and scores. Arata said he felt nervous with other examinees sitting near him, close enough for them to hear him. Yamato also hesitated to articulate his speeches because doing so could disturb the examinees sitting close to him. Gaku could not concentrate well because a person nearby was not wearing a headset and spoke loudly. In terms of technical issues, Takeru’s microphone could not pick up his voice during the microphone check. It turned out that his voice was too soft for the microphone to detect; however, the “failure” of the microphone check made him nervous during the S section.

Listening (L)

Subaru and Gaku pointed out that the formats of the TOEFL iBT L section were dissimilar from those of the Center Test, which made it difficult to score higher in the former. Both Subaru and Gaku said that listening was much faster and longer on the iBT than on the Center Test. Another challenging factor was that, in the iBT, answer choices were not shown until the end of the speech. In the Center Test and the TOEFL Institutional Testing Program (ITP), both of which the participants had taken, answer choices were printed on a test book, so they could read them before the speech started to get an idea of what it would be about. Gaku said that, during the iBT, speeches are read only once, whereas during the Center Test, they are read twice. He explained that he would often translate the English text into Japanese when listening to (and even reading) English texts; thus, he could not catch up with the fast, long speeches while translating in the iBT L section. In contrast, he was able to catch up with L texts and managed to answer questions in the Center Test due to the repetition of the speech (see He & Jiang, 2020; Pusey, 2020, for previous studies).

The order of the skill sections may also have slightly affected the performance. Takeru mentioned that in an R-and-then-L integrated task, he was not able to concentrate well on L and used too many cognitive resources to complete the R section.

Writing (W)

Only one point was raised regarding the test formats used in the W section. Takeru could not type fast, which slowed down his answering process in the W section. Although Japanese students are familiar with texting on smartphones using an on-screen keyboard, they are typically less familiar with conventional physical keyboards. At the university, they wrote essays using a laptop. However, some learners like Takeru may have needed more typing practice. Additionally, for TOEFL iBT, a standard English-language (QWERTY) computer keyboard was used (https://www.etsglobal.org/hu/en/test-type-family/toefl-ibt-test), which has a slightly different layout from that of a Japanese keyboard. Thus, typing speed and keyboard layout could have affected the learners’ W performance (see Ling, 2017, for previous studies).

Conclusion

The current study examined characteristics of the four-skill profiles that are frequently observed across datasets (RQ1 in Study 1 [via a quantitative approach]) and the learners’ perceived reasons for having uneven profiles (RQ2 in Studies 2 and 3 [via a quantitative and qualitative approach]) among Japanese English learners. The findings are summarized in Table 12. Study 1 used 10 datasets from five standardized four-skill tests to search for general patterns in profiles. We found 75 various types of skill profiles, and learners with uneven profiles were more common (over 30%) than those with a flat profile. Skill profiles frequently observed across databases included LSW<R (in eight datasets), SW<LR (in seven datasets), S<LRW (in seven datasets), and flat (in four datasets). A one-level difference across skills accounted for the highest percentage (up to 70.00%), followed by a two-level difference (up to 40.16%) and a flat profile (up to 29.00%).

Table 12 Joint display of the current study’s results

Based on a quantitative approach using a questionnaire, Study 2 examined why some learners had uneven profiles (RQ2). Those with substantially uneven skill profiles reported that they were poor at the skill (e.g., up to 68.18%); they had not studied the skill sufficiently (up to 52.78%), and they did not know the test formats (up to 24.32%) or were tense (up to 36.36%).

Study 3 further examined RQ2 by qualitatively analyzing responses from interviews with six learners exhibiting uneven skill profiles. Their responses revealed that their skill imbalances can be explained by three main factors: (a) insufficient practice, (b) particular subskills or processes required to accomplish test tasks, and (c) unfamiliarity with test formats and test-taking environments. There were also underlying factors behind these factors, such as the impact of entrance examinations on teaching and learning, and the difficulty in understanding a long listening text played only once. Study 3’s qualitative findings supported Study 2’s quantitative findings, because (a), (b), and (c) as found in Study 3 were three of the four key reasons reported in Study 2.

The current study reinforced the importance of considering uneven profiles of L2 proficiency, as suggested by previous studies (e.g., North, 2021; Hulstijn, 2015). However, we found far more uneven profiles (i.e., 75 profiles in Study 1) than those reported in previous studies (e.g., Ginther & Yan, 2018, showing three profiles) as we computed the percentage of each skill profile to examine detailed skill profiles. We also reported more detailed explanations of learners’ self-perceived reasons for having uneven profiles (in Studies 2 and 3, e.g., insufficient practice) than previous studies that suggested that extensive test preparation is only one of the factors that produce them (e.g., Ginther & Yan, 2018).

In future studies, in addition to expanding learner groups and tests to examine the generalizability of the current findings, we suggest examining how skill profiles affect L2 performance in general, academic, and other specific contexts. In particular, learners with uneven profiles of SW<LR and SW<<LR should be examined along with their L2 test or real-life performance. These two-skill profiles were identified frequently in the current study and were considered problematic by Ginther and Yan (2018), and Bridgeman et al. (2016) as learners with these profiles performed more poorly than expected in academic contexts. It remains to be examined whether and how those learners could have compensated for their weak skills by effectively using their strong skills (Harsch, 2014; Hulstijn, 2015). There might be minimally required degrees of skill or an optimal balance of four skills for learners to perform effectively. These insights would help explore how students with extremely uneven profiles can be taught and how this information can be integrated into score reports.

As for the implications of our study, we argue that effective use of skill profiles in score reports would help test developers, admission and placement officers, learners, teachers, and other test users understand learners’ strengths and weaknesses. Test developers can utilize these results to consider how to present skill profiles and convey the interpretations in score reports and supplementary materials provided to test-takers and users to enhance the usefulness of their tests. For example, they can flag two- or more-level differences between skills on the score reports, so that test users can consider them when making decisions. Those who are involved in admission and placement can also incorporate the use of both total and skill scores, especially scores of important skills (Ginther & Yan, 2018), or pass on the skill profile information and recommendations to teachers and administrators who may plan remedial instructions.

When test-takers are shocked to see their uneven profiles in the score reports, we suggest that a skill imbalance, specifically a skill difference of one CEFR level, is not surprising. For those who had two- or more-level differences, we suggest that they fully utilize the score report to understand their strengths and weaknesses, and that they review possible factors affecting the uneven profiles, such as insufficient practice, particular subskills or processes required to accomplish test tasks, and unfamiliarity with test formats and test-taking environments. To fully demonstrate L2 skills and avoid obtaining uneven skill profiles due to the lack of test format knowledge, it may be necessary to do prior familiarization practice with the test format, for example, by watching videos (e.g., https://www.youtube.com/user/TOEFLtv). Furthermore, when the uneven profiles come from true skill imbalance, learners and teachers may need to decide whether to focus on weaker or stronger skills for further study based on their contexts and needs. While one direction is to improve a weaker skill, a stronger skill could be further improved to compensate for the weaker one. Teachers can also make informed decisions on how to provide students with useful feedback, remedial instructions for lower skills, and training for strengthening higher skills, according to the learner’s needs. These activities are in line with the global trends in learning-oriented assessment (Gebril, 2021) by relating assessment results with instruction and learning.

Availability of data and materials

The datasets for the current study are available from the corresponding author on reasonable request.

Notes

  1. To facilitate understanding, = and the spaces before and after < and = were omitted.

Abbreviations

CBT:

Computer-based testing

CEFR:

Common European Framework of Reference for Languages

Center Test:

National Center Test for University Admissions

EFL:

English as a foreign language

GPA:

Grade point average

GRE:

Graduate record examination

GTEC:

Global Test of English Communication

iBT:

Internet-based test

L:

Listening

MEXT:

Japan’s Ministry of Education, Culture, Sports, Science and Technology

R:

Reading

RQ:

Research question

S:

Speaking

SEM:

Standard error of measurement

TEAP:

Test of English for Academic Purposes

TOEFL:

Test of English as a Foreign Language®

TOEFL Junior:

TOEFL Junior® Comprehensive

TOEIC:

Test of English for International Communication®

W:

Writing

References

Download references

Acknowledgements

We would like to thank the people and institutions that supported us in Study 1: Yuko Kashimada, Takuya Akimoto, and Yumiko Tashiro for Dataset 1; the Global Communication & Testing for Datasets 2 and 3; Keita Nakamura and Eiken Foundation of Japan for Datasets 4 and 5; Shin’ichiro Ishikawa, Mineo Mitsuhashi, Institute for International Business Communication, and Yuka Ishikawa for Dataset 7; and Ryoko Fujita and Yumi Koyamada for Dataset 8.

Funding

This work was supported by the Japan Society for the Promotion of Science (JSPS) KAKENHI, Grant-in-Aid for Scientific Research (C), Grant Numbers 17K02983 and 26370737, and the Juntendo University General Education Joint Research Grants 2015 and 2016.

Author information

Authors and Affiliations

Authors

Contributions

RK and YI planned Study 1 and collected the datasets. The first three authors planned Studies 2 and 3 and collected and transcribed the data. All the authors analyzed the data. RK, TA, and YI wrote the initial draft, which was extensively revised, and the final manuscript was approved by all the authors.

Corresponding author

Correspondence to Rie Koizumi.

Ethics declarations

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

40468_2022_203_MOESM1_ESM.docx

Additional file 1. Questionnaire regarding the TOEFL iBT (originally in Japanese; translated into English for presentation purposes).

Appendix

Appendix

Table 13

Table 13 All skill profile types that appeared in Study 1 (%)

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Koizumi, R., Agawa, T., Asano, K. et al. Skill profiles of Japanese English learners and reasons for uneven patterns. Lang Test Asia 12, 53 (2022). https://doi.org/10.1186/s40468-022-00203-3

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s40468-022-00203-3

Keywords

  • Japanese learners of English
  • Second-language English proficiency
  • Listening
  • Reading
  • Score report
  • Skill imbalance
  • Skill profiles
  • Speaking
  • Writing