The Concise Encyclopedia of Applied Linguistics. Carol A. Chapelle

Чтение книги онлайн.

Читать онлайн книгу The Concise Encyclopedia of Applied Linguistics - Carol A. Chapelle страница 80

The Concise Encyclopedia of Applied Linguistics - Carol A. Chapelle

Скачать книгу

best ensure that the test provides a valid indication of learners' listening abilities.

      Developments in computer technology expand the potential for different types of listening input and ways of determining comprehension of input. For instance, technology allows acoustic signals to be easily accompanied by various types of input such as visual stimuli, which can make tasks more realistic.

      Computers also increase the potential for using test items that require short constructed responses. Developers create model answers and identify the key words and phrases in them, and then these key words and phrases, along with acceptable synonyms, can be used to create a scoring algorithm. Reponses to items that contain part or all of the targeted information can be given partial or full credit (Carr, 2014).

      Computer technology may, however, have a negative effect on the assessment of listening. The tendency to use technology because it is available and attractive may lead to assessments that do not validly measure listening comprehension (Ockey, 2009). For instance, including attractively colored interfaces or interesting sounds, which are not part of the message to be comprehended, may be distracting to some test takers. Such distractions could lead to invalid scores for affected individuals. Moreover, computer scoring systems cannot make human judgments that may be important when assessing language (Condon, 2006). For instance, in a summary task scored by a computer, a test taker who paraphrases may be assigned a low score because the computer does not recognize that the test taker has comprehended the input; the scoring algorithm only assigns points for vocabulary (or common synonyms of the vocabulary) found in the input. It is therefore important that test developers who use computers to assess listening recognize the strengths and limitations of this technology.

      An increasing number of researchers support listening assessments which include as much contextual information in the input as possible (Gruba, 1999; Buck, 2001; Ockey & Wagner, 2018). Of particular interest to most of these researchers is that the acoustic signal be accompanied by associated visual stimuli, such as a video which shows the speaker.

      Research which has compared test takers' scores on audio‐only assessments with tests that include visual stimuli as well as the audio input have produced mixed results. Some studies have found that including visuals leads to increased scores (Shin, 1998; Sueyoshi & Hardison, 2005; Wagner, 2010), while others have failed to find a difference in scores for the two conditions (Gruba, 1989; Coniam, 2000). A recent study by Batty (2018) may help to explain these contradictory findings. He found that particular question types are impacted in different ways by including visuals. His research indicated that implicit items were made much easier by visuals while explicit items were less affected by including visual stimuli. It may be that studies which had mostly explicit items failed to find a difference between the audio‐only and audio accompanied by visual information.

      Eye‐tracking research has also provided increased understanding of listening processes while test takers attempt to comprehend listening input. Using eye‐tracking technology, Suvorov (2015) considered dwell time (how long eye gaze fixates on a particular visual stimuli) and found that test takers paid more attention to content videos than context videos. Also using dwell time with eye‐tracking technology, Batty (2016) found that test takers spent over 80% of their time observing facial cues when watching videos.

      Researchers increasingly argue that the aim of assessing listening should not necessarily be to attempt to isolate comprehension from other language abilities. These researchers contend that listening is commonly interactive, meaning most listening includes opportunities to ask for clarification and that listeners are typically expected to respond after listening (Douglas, 1997; Ockey & Wagner, 2018). Other research indicates that listening and speaking cannot be separated in interactive discussions among two or more individuals (Ducasse & Brown, 2009; Galaczi, 2014). As a result of these conceptualizations of listening and research findings, test developers have begun to create listen–speak tasks (and other integrated listening items), which require both listening and speaking. They contend that it may not be appropriate or even possible to measure listening as distinct from oral production in an interactive communication context. Such an approach limits concerns about measuring more than “listening” with listening assessments.

      SEE ALSO: Assessment of Integrated Skills; Uses of Language Assessments; Validation of Language Assessments

      1 Bachman, L. F., & Palmer, A. S. (2010). Language assessment in the real world. Oxford, England: Oxford University Press.

      2 Batty, A. O. (2016). The impact of visual cues on item response in video‐mediated tests of foreign language listening comprehension (Unpublished doctoral thesis). Lancaster University, England.

      3 Batty, A. O. (2018). Investigating the impact of nonverbal communication cues on listening items. In G. J. Ockey & E. Wagner, Assessing L2 listening: Moving towards authenticity (pp. 161–85). Philadelphia, PA: John Benjamins.

      4 Brindley, G., & Slatyer, H. (2002). Exploring task difficulty in ESL listening assessment. Language Testing, 19(4), 369–94.

      5 Buck, G. (2001). Assessing listening. Cambridge, England: Cambridge University Press.

      6 Buck, G. (2018). Preface. In G. J. Ockey & E. Wagner, Assessing L2 listening: Moving towards authenticity (pp. xi–xvi). Philadelphia, PA: John Benjamins.

      7 Carr, N. T. (2014). Computer‐automated scoring of written responses. In A. Kunnan (Ed.), The companion to language assessment (Vol. 2, Pt. 8, p. 64). Malden, MA: John Wiley.

      8 Condon, W. (2006). Why less is not more: What we lose by letting a computer score writing samples. In P. Ericsson & R. Haswell (Eds.), Machine scoring of student essays: Truth and consequences (pp. 209–20). Logan: Utah State University Press.

      9 Coniam, D. (2000). The use of audio or video comprehension as an assessment instrument in the certification of English language teachers: A case study. System, 29, 1–14.

      10 Douglas, D. (1997). Testing speaking ability in academic contexts: Theoretical considerations. TOEFL monograph series, 8. Princeton, NJ: Educational Testing Service.

      11 Ducasse, A. M., & Brown, A. (2009). Assessing paired orals: Raters' orientation to interaction. Language Testing, 26(3), 423–43.

      12 Dunkel, P., & Davis, A. (1994). The effects of rhetorical signaling cues on the recall of English lecture information by speakers

Скачать книгу