EFL university teachers’ perceptions of writing assessment training

Percepciones de Docentes Universitarios sobre un Taller de Evaluación de Escritura en

Inglés como Lengua Extranjera

Elsa Fernanda González[1]

 

ABSTRACT

Assessment literacy is a term that has arisen from the worldwide constant use of assessment data and the need to help teachers understand and apply assessment procedures in their language classrooms (Malone, 2013; Inbar-Lourie, 2013). It involves the theoretical knowledge of assessment, its principles and the know how to (Fulcher, 2012) apply them in each specific context. Specifically, the assessment of writing remains a difficult activity that English as a Foreign Language (EFL) Mexican teachers are required to conduct as a regular activity of their language teaching profession. However, these activities are carried out, most of the times, without the proper training, guidance and consideration of teachers’ needs to assure students’ assessment validity and reliability. The study explores the perceptions that 48 Mexican EFL university teachers had in relation to the effectiveness of two writing assessment training sessions provided to them in a period of twelve months. Data obtained from a background questionnaire and an online post training questionnaire suggested that half of the teacher participants did not have previous writing assessment training nor for the use of scoring tools such as analytic and holistic rubrics. Additionally, it was found that although teachers found the sessions useful and practical for their future assessment practice they considered more practice using assessment rubrics and understanding the writing assessment process was needed. Teachers’ perceptions are also analyzed regarding the perceived changes that training encouraged. It is concluded that the inexperience with writing assessment that most of the teachers stated to have may have influenced the perceptions participants reported. Implications for the language student, teacher and institution are discussed in the conclusions of the paper.

 

Palabras clave: assessment literacy, EFL writing assessment, EFL teachers, teacher training, scoring rubrics.

 

RESUMEN

La alfabetización de evaluación es un término que ha surgido del uso constante en el ámbito internacional de los datos de evaluación y la necesidad de ayudar a los maestros a comprender y aplicar los procedimientos de evaluación en sus aulas de idiomas (Malone, 2013; Inbar-Lourie, 2013). Implica el conocimiento teórico de la evaluación, sus principios y el saber hacer (Fulcher, 2012) que se aplican en cada contexto específico. Específicamente, la evaluación de la escritura sigue siendo una actividad difícil que los profesores mexicanos de inglés como idioma extranjero (EFL) deben realizar como una actividad regular de su profesión de enseñanza de idiomas. Sin embargo, estas actividades se llevan a cabo, la mayoría de las veces, sin la capacitación, orientación y consideración adecuada de las necesidades de los maestro para asegurar la validez y confiabilidad de la evaluación de los estudiantes. Considerando esta problemática, el presente estudio explora las percepciones que 48 profesores universitarios mexicanos de inglés como lengua extranjera tenían en relación a la efectividad de dos sesiones de capacitación de la evaluación de escritura que se les brindaron en un  período de doce meses. Los datos obtenidos de un cuestionario de antecedentes y un cuestionario electrónico posterior a la capacitación sugirieron que la mitad de los docentes participantes no tenían una capacitación previa en evaluación de escritura ni para el uso de herramientas de evaluación como las rúbricas analíticas y holísticas. Además, se encontró que a pesar de que los profesores consideraban que las sesiones eran útiles y prácticas para su futura práctica de evaluación, consideraban que era necesario comprender el proceso de evaluación de la escritura. Las percepciones de los maestros también se analizan con respecto a los cambios percibidos que la capacitación alentó. Se concluye que la inexperiencia con la evaluación de la escritura que la mayoría de los maestros tenia pudo haber influido en las percepciones manifestadas por los participantes. Las  implicaciones para el estudiante de idiomas, el maestro y la institución se discuten en las conclusiones del presente estudio.

 

Keywords: alfabetización evaluativa, evaluación de la escritura en inglés como lengua extranjera, docentes de inglés como lengua extranjera, capacitación docente, rubricas de evaluación.

 

INTRODUCTION

In many higher education institutions of Mexico, English as Foreign Language (EFL) teachers are required to teach and assess the four language skills on a regular basis. They need to select an assessment method; develop the assessment tool or use one provided by the program manager; administer and score the tool; interpret and make decisions related to the score; communicate the results and cope with the consequences that assessment and evaluation may have (Crusan, 2014; Fulcher, 2012; Stoynoff and Coomb, 2012; Weigle, 2007). To perform all these activities university language teachers need to be assessment literate. 

The lack of assessment literacy may not only result in a heavier workload for teachers; it may also negatively affect the validity and reliability of the assessment of their students’ writing abilities.

 

To develop writing assessment literacy teachers require continuous, well-planned training. Lack of training often results in teachers’ uneasiness and distrust in their abilities to assess their students’ written work (González, 2017). Training may favor score and assessment reliability and consistency (Bachman and Palmer, 2010; Hamp-Lyons, 2003; Weigle, 2007). Teachers, however, might not value the training received, or their views of training may impede a positive impact on their assessment practices. Teachers’ perceptions of writing assessment training are therefore, a legitimate field of inquiry.

 

Assessment literacy means being familiar with and using measurement practices to assess the language used by students for a class (Malone, 2013). Assessment literacy research began in the late 1990s and it has investigated writing teachers’ assessment training needs; teachers’ perceptions of assessment training; and the impact of trainers’ backgrounds on the content and procedures of the assessment training they provide (Bailey and Brown, 1996, 2008; Fulcher, 2012; Hasselgreen and col. 2004; Jeong, 2013; Nier  and col. 2013; Stiggins, 1995). Studies on assessment literacy take place in classrooms of English as a first language (L1), English as a second language (ESL), and English as a foreign language (EFL). Studies that examine assessment training in ESL and EFL contexts, focus mainly on the impact of raters’ training; raters’ backgrounds; raters’ use of rubrics; raters’ gender and other issues of large-scale testing (Barkaoui, 2007, 2011; Eckes, 2008; Esfandiari and Myford, 2013; Lim, 2011).

 

Specifically in EFL education, Nier and col. (2013), focused on analyzing a blended learning assessment course and its usefulness to participants. They administered a post-training questionnaire to 35 teachers and analyzed the group discussions conducted during the face-to-face encounters. Results indicate that most participants considered the blended learning approach as useful, but required more examples to understand the processes of assessment. Participants identified the course and mode of course as a helpful and useful mode of professional development.

 

In another study, Jeong (2013) examined teacher trainers’ understanding of assessment and the ways in which their assessment background influenced the outcomes of their assessment courses. Participants were 140 instructors of language assessment courses (both language testers and non-language testers).  Data were collected with the use of an online survey and a telephone interview. Findings show that there were significant differences in the content of the courses depending on the instructors’ background in six topic areas: test specifications, test theory, basic statistics, classroom assessment, rubric development, and test accommodation. Non-language testers were less confident in teaching technical assessment skills compared to language testers and had a tendency to focus more on classroom assessment issues. The researcher recommends language testers to share their knowledge and make it accessible to those who are part of the language assessment community. Research still needs to explore the writing assessment literacy of EFL teachers; the ways in which EFL teachers assess writing; and the impact of EFL teachers’ perspectives of assessment on their assessment practices. Research should also explore the assessment context, the assessment needs, and the perceptions of assessment training of EFL teachers in Latin American countries.

 

In Mexico, undergraduate students in most universities are required at least a B1 level of proficiency in a language other than Spanish. Therefore, teachers need to be assessment knowledgeable; have practical assessment skills; have the capacity to connect classroom assessment to large-scale tests; and maintain the focus on students’ learning as the main purpose of assessment. Assessment literacy is particularly important to develop the complex ability of EFL writing.

 

This study examined the perceptions of teachers that participated in a two-session writing assessment workshop in the 2014-2015 school calendar. The research questions addressed were the following:

 

1)     What are the teachers’ perceptions of the usefulness of the writing assessment training received?

2)     How do writing assessment training influence classroom assessment practices?

 

METHODOLOGY

This study uses a cross-sectional, non-experimental, intervention design. It is descriptive and exploratory. It does not intend to generalize the results to other populations. Instead, its purpose is to analyze the unique traits that characterize the small group of participants. Data collection and analysis were driven by a mixed-methods approach. Combining quantitative and qualitative data allowed a better understanding of the teachers’ perceptions of assessment and the assessment training they received (Creswell & Plano Clark, 2011; Creswell, 2013). In an effort to care for the validity and reliability of the findings  here portrayed data was shared with an experienced researcher in the area of applied linguistics following a peer checking process (Dörnyei, 2007) that allowed the comparison of results obtained from both researchers. To diminish the Hawthorne Effect as well as the Social Desirability Bias (Dörnyei, 2007) effect a data triangulation method was conducted during which specific data was elicited in different forms and structures within the same online questionnaire.

 

Participants and research context

Participants were teachers of three universities (19 participants) and one language institute (29 participants) in the northeastern corner of Mexico. Initially, 150 teachers were invited to take part in the study, since they were in service teachers at the time of the study, and teaching in university settings. However, only 48 gave their informed consent to participate. A convenience sampling method was used (Dörnyei, 2007) which emphasizes the inclusion of those participants who were available and willing to take part in the study.

 

The teachers’ institutions of affiliation used different programs and methods to teach and assess writing. Regarding their assessment policies, all four schools required their teachers to calculate a holistic score (0-10 or 0-100) that integrated students’ EFL writing proficiency with other language skills. Teachers from the language institute reported that the institution had established writing tasks and scoring rubrics to assess the writing abilities of students. University teachers, on the other hand, stated that their institutions did not provide assessment guidelines and they were free to decide on their assessment approaches. Neither the language institute nor the universities gave their teachers EFL writing assessment training.

 

Writing assessment training

The researcher, who was also the trainer, delivered the workshop in two sessions. The first session focused on the nature of EFL writing, writing assessment, and the use of holistic and analytic rubrics. The second session focused on the importance of using rubrics to assess writing skills and giving trainees time for practice. Teachers reflected on the characteristics of their teaching contexts and their assessment practices. The trainer structured the workshop in accordance with the concept of “assessing for learning” (Stiggins, 1995); the manual for language examinations (Concil of Europe, 2002; 2009a; 2009b) of the Common European Framework of Reference (CEFR); and the manual for language test development and examining (2011) of the Association of Language Testers in Europe (ALTE). The assessment training given to participants included: (a) guided discussion of previously scored samples; (b) independent marking and follow-up discussion of scores; and (c) independent marking and pair discussion of scores.

 

Data collection instruments

Two data collection instruments were used: a background questionnaire and an online post-training questionnaire. The background questionnaire was paper-based and administered on-site during the first training session. It gathered information about the participants’ EFL teaching and assessment experience.

 

A combination of eight multiple-choice and three open-ended questions were included in the background questionnaire to provide informants with opportunities for free expression (Nunan, 1992). The post-training questionnaire was written in the participants’ L1 (Spanish) and delivered electronically with the use of a survey generation and research platform for members of the University of Southampton (found at: https://www.isurvey.soton.ac.uk/). The tool made thedata collection processes effective for the researcher and attractive for the participants (Dörnyei & Taguchi, 2010). The survey included Likert-scale items, closed and open questions (Dörnyei, 2007).  It was pilot tested (Dörnyei, 2003) with a group of EFL teachers that were not part of this study.

 

Data collection and analysis procedures

Data collection for the study involved two stages in a period of twelve months. In the first stage, the researcher asked participants to complete the background questionnaire and delivered the first training session, which lasted approximately three hours which included a 20-minute break. Eight months later, in stage two, the researcher provided the second assessment training session, which took approximately two to three hours to complete. During the session teachers engaged in assessment practice with scoring rubrics, group discussion and benchmark scoring of sample papers. They also shared their reflections on the changes they had observed in their assessment practice, after receiving the first assessment training session. Then, the researcher explained how to answer the online post-training questionnaire and notified teachers they would receive a link by email to answer it. Teachers answered the questionnaire two to three weeks after completing the second training session.

 

For the analysis, descriptive statistics was used with data that came from the closed-ended items of both questionnaires. The information was introduced into the Statistical Package for the Social Sciences (SPSS. 23) to calculate Means, Mode, and frequency of data items. The sample of teachers did not allow for inferential statistics analysis and therefore, only exploratory and descriptive statistics were used.

 

For the analysis of open-ended questions, themes were identified and clustered into categories. Each category was given a code and frequencies for each code were calculated (Creswell, 2015).Participants’ responses were analyzed in Spanish to avoid translation of data bias or subjectivity (Pavlenko, 2007). Once the analysis ended, responses were translated into English to report the results.

 

RESULTS

Participants’ background Participants were 65 % female and 35 % male EFL teachers. Their ages ranged from 20 to 52 years. Regarding their teaching experience, 67 % had taught for five or less years; 25 % had five to nine years of experience; and only 8 % had been teachers for 10 or more years.  As to their professional background, 38 % were undergraduate students working as English language teachers; 29 % had under- graduate or graduate degree and a teaching certificate (Teaching Knowledge Test or the In-Service Certificate of Language Teaching by Cambridge English Language Assessment). Finally, 33 % of the trainees had undergraduate or graduate degree and lacked a teaching certificate. Information on the participants’ background is shown on Table 1.

 

 

                                                                                                                                                            

 

 

 

 

To compare the participants’ assessment practices before and after the training, the background questionnaire investigated their previous assessment training and use of assessment tools. As shown on Table 1 above, 54 % of the participants responded that they had not received assessment training, while 46 % answered that they had received assessment preparation. Trainees also reported the assessment frequency in their teaching of EFL writing. Of the 48 teachers, 36 % reported that they often assessed writing; 34 % responded that they assessed writing always. Together, people that often and always assessed writing, made 70 % of the sample. However, 20 % sometimes assessed their students; 5 % never did; and 5 % rarely assessed writing in their classrooms.

 

In relation to assessment tools, teachers were asked about their rubric training and rubric use. Table 1 shows that 56 % responded that they had never received training and 44 % stated that they had received preparation in the use of assessment tools. This is consistent with 44 % of the teachers that informed that they used rubrics (23 % stated that they always used rubrics and 21 % responded that they often did), and 56 % that reported an infrequent use of rubrics (17 %, rarely; 15 %, sometimes; 15 %, never; and 9 %, hardly ever).

 

What are the teachers’ perceptions of the writing assessment training received? In general, the EFL writing assessment workshop was well accepted by the trainees. Most of them (90 %) considered that the content was clear and understandable. A high percentage of them (96 %) either agreed or strongly agreed with the idea that the training was practical in their subsequent assessment practice.  The majority of them (92 %) agreed or strongly agreed with the item that read: The information and practice shared is useful for future writing assessment. These results are shown on Table 2.

 

 

How do writing assessment training influence classroom assessment practices?

As can be portrayed on Table 3, the he teachers’ perception of the influence of the training on their assessment practice was generally positive. A large group (88 %) agreed or strongly agreed with the notion that scoring their students’ pieces of writing became easier for them after the workshop. Many of them also perceived that scoring (90 %) and the use of rubrics (90 %) became more efficient.

 

The use of rubrics became easier (90 %) and so they considered that the rubrics provided by the trainer during the workshop would be useful in their subsequent assessment practice. However, 13 % of the participants did not plan to use a scoring tool to assess the writings of their EFL students.

 

The open-ended questions related to the perceived changes in the assessment practices of teachers as a result of the writing assessment training, revealed three major themes.  The themes came from those teachers that: (1) perceived their assessment as more objective after receiving the training; (2) those that considered their assessment became more efficient in terms of speed and practicality; and (3) those teachers that did not perceive any change in their assessment practices as a result of taking the EFL writing assessment workshop. The open-ended questions related to the perceived changes in the assessment practices of teachers as a result of the writing assessment training, revealed three major themes.  The themes came from those teachers  that: (1) perceived their assessment as more objective after receiving the training; (2) those that considered their assessment became more efficient in terms of speed and practicality; and (3) those teachers that did not perceive any change in their assessment practices as a result of taking the EFL writing assessment workshop. As to the reasons for considering that the use of rubrics made their assessments more efficient, they affirmed that the perceived that their scoring became more impartial.

 

…the rubrics provided in the workshop are useful to supplement the rubrics we already used and to make a more objective assessment (TP04).

 

Another participant that considered that the workshop contributed to a more objective view expressed the following,

using rubrics to evaluate writing changed because I managed to understand that when assessing a text, I must take into account several things, not only spelling or grammatical errors.  I also learned that with a rubric it is easier for both, the teacher and the student, to be clear about the features of writing that will be assessed and to ensure that the score awarded is reliable” (TP302).

 

The second theme emerged from the participants who considered their assessment became more efficient referred to the time invested in assessing students work. The following extract of a trainee’s written comments illustrates this view.

 

The use of rubrics has notably facilitated me the assessment of students’ writing; it is a facilitating tool and it saves time (TP31).

 

Another teacher considered that after the training, his assessment became more precise. The following comment reflects this view.

It’s easier for me to differentiate if a student belongs to a specific grade of competence described in the rubric, without hesitating or doubting when giving the score (TP36 ).

 

Most trainees that perceived no change in their assessment practices after receiving the training reported that the rubrics provided in the workshop were very similar to those they were already using in workplaces (TP14).

 

One teacher perceived that the training received was more useful to analyze his own use of rubrics than to assess his students’ EFL writing (TP35).

 

DISCUSSION

This study analyzed the perceptions that 48 EFL Mexican university teachers regarding a writing assessment workshop. It also examined the teachers’ perceived changes in their assessment practices after attending the writing assessment training. Findings indicate that they perceived the training as useful and practical. They also considered the training resulted in a more efficient and objective assessment practice, as well as an easier and less time-consuming scoring of students’ written work. Interestingly, a large part of the sample considered implementing scoring tools in their classroom assessment after the training, thus making a change in their assessment practices. These results seem to echo those of Nier, Donnovan and Malone (2013) in which 35 language teachers answered a post-training questionnaire. Still, four teachers disagreed and strongly disagreed with the statement that training was useful for their assessment practice. They also perceived that training did not change their assessment practice. However, this study focused on their perceptions of their practice and not on what they in fact do in the classroom.

 

Future research could focus on the concrete assessment processes that they make happen in the classroom, to give them a better assessment preparation. The majority of the EFL teachers who took the assessment workshop seemed to be conscious of their language assessment weaknesses. They were always willing to participate in pair and class discussion, and to practice the use of holistic and analytic rubrics. However, a small group of teachers seemed to refuse the use of scoring tools in their writing classes.

 

This finding seems to be related to their specific teaching experience and professional backgrounds. The participants of this study, as foreign language teachers worldwide, come from different professional fields, which may influence their understanding of assessment. Jeong (2013), found significant differences in the content of assessment training courses depending on the instructors’ background.

 

The influence of language assessment trainees on the ways they perceive assessment would need further research. Finally, this study involved two data-collection instruments that involve indirect contact with participants and favor short responses. Future studies could consider the use of other collection instruments, such as face-to-face or stimulus –recall interviews, which allow direct contact and more nourished responses.

 

CONCLUSIONS

The conclusions that emanate from this study highlight the importance that writing assessment training represents for the language student, the language teacher and for any type of language institution. On many occasions, the future of a language student is determined by scores provided by a teacher in the classroom or on a large-scale test. Teachers in the Mexican EFL context on the other hand, are required to assess language skills on a regular day-to-day basis. Therefore, it seems rather unfair for the student and the teacher to conduct these assessments without prior and proper training, jeopardizing the validity and reliability of assessment and the students’ future academic life. On the other hand, the results of this study could also serve teacher trainers and language program managers or coordinators to understand the needs of their teachers and their views in terms of writing assessment. This with the purpose of comparing and contrasting them with the institutions’ teaching goals and teacher training possibilities so that appropriate sessions are provided, making the breach between the teacher and assessment literacy as small as possible.

 

REFERENCES

Bailey, K. M. and Brown, J. D. (1996). Language testing courses: What are they? In A. Cumming and R. Berwick (Eds.), Validation in language testing (pp. 236–256). UK: Multilingual Matters.

Bailey, K. M. and Brown, J. D. (2008). Language testing courses: What are they in 2007? Language Testing. 25(3): 349-383.

Bachman, L.F. and Palmer, A. (2010). Language Assessment in Practice. Oxford, U.K.: Oxford University Press.

Barkaoui, K. (2007). Rating scale impact on EFL essay marking: A mixed-method study. Assessing Writing. 12(2): 86–107.

Barkaoui, K., (2011). Effects of marking method and rater experience on ESL essay scores and rater performance. Assessment in Education: Principles, Policy & Practice. 18(3): 279–293.

Creswell, J. W. & Plano Clark, V. L. (2011). Designing and conducting mixed methods research. Thousand Oaks, CA: Sage Publications.

Creswell, J. W. (2013). Research design: Qualitative, quantitative, and mixed methods approaches. Thousand Oaks, CA: Sage Publications.

Creswell, J.W. (2015). A Concise Introduction to Mixed Methods Research. Thousand Oaks, California, USA: Sage Publications.

Crusan, D. (2014). Assessing writing. In Anthony John Kunan (Ed.) The companion to language assessment (pp. 206-217). UK: John Wiley & Sons.

Council of Europe (2002). Common European Framework of Reference for Languages: Learning, Teaching and Assessment. Strasbourgh, France: Council of Europe.

Council of Europe (2009a). The manual for language test development and examination. Strasbourgh, France: Council of Europe.

Council of Europe (2009b). Manual for relating language examinations to the Common European Framework of Reference for Languages: Learning, teaching and assessment. Strasbourgh, FR: Council of Europe.

Dörnyei, Z. (2003). Questionnaires in second language research: Construction, administration, and processing. Mahwah, NJ: Lawrence Erlbaum Associates.

Dörnyei, Z. (2007). Research methods in applied linguistics: Quantitative, qualitative and mixed methodologies. Oxford: Oxford University Press.

Dörnyei, Z. and Taguchi, T. (2010). Questionnaires in second language research: Construction, administration and processing. NY: Routledge.

Eckes, T. (2008). Rater types in writing performance assessments: A classification approach to rater variability. Language Testing. 25(2): 155–185.

Esfandiari, R., Myford, C. (2013). Severity differences among self-assessors, peer-assessors, and teacher assessors rating EFL essays. Assessing Writing. 18(2): 111–131.

Fulcher, G. (2012). Assessment Literacy for the Language Classroom, Language Assessment Quarterly. 9(2): 113-132.

González, E.F. (2017). The challenge of EFL writing assessment in Mexican Higher Education. En Higher Education English Language Teaching and Research in Mexico (p. Grounds & C. Moore, Eds.) (pp.73-100). British Council: Mexico City.

Hamp-Lyons, L. (2003). Writing teachers as assessors of writing. In Kroll, B. (Ed.) Exploring the Dynamics of Second Language Writing (pp. 162-189). New York, USA: Cambridge University Press.

Hasselgreen, A., Carlsen, C., and Helness, H. (2004). European survey of language testing and assessment needs. Part 1: General findings. Gothenburg, Sweden: European Association for Language Testing and Assessment. Retrieved from. [Online]. Available in: http://www.ealta.eu.org/documents/resources/survey-report-pt1.pdf.

Jeong, H. (2013). Defining assessment literacy: Is it different for language testers and non-language testers? Language Testing. 30(3): 345-362.

Lim, G. (2011). The development and maintenance of rating quality in performance writing assessment: A longitudinal study of new and experienced raters. Language Testing. (28): 543-560.

Malone, M. E. (2013). The essentials of assessment literacy: Contrasts between testers and users. Language Testing. 30(3): 329-344.

Nier, V. C., Donovan, A. E., and Malone, M. E. (2009). Increasing assessment literacy among LCTL instructors through blended learning. Journal of the National Council of Less Commonly Taught Languages. (9): 105–136.

Nunan, D. (1992). Research methods in language learning. NY: Cambridge University Press.

Pavlenko, A. (2007). Autobiographic narratives as data in applied linguistics. Applied Linguistics. 28(2): 166-188.

Stiggins, R. J. (1995). Assessment literacy for the 21st century. Phi Delta Kappan. 77(3): 238-245.

Stoynoff, S. and Coombe, C. (2012). Professional development in language assessment. In Christine Coombe, Peter Davidson, Barry O’Sullivan, Stephen Stoynoff (Eds.) The Cambridge guide to second language assessment (pp. 122-130). NY: Cambridge University Press.

Weigle, S. C. (2007). Teaching writing teachers about assessment. Journal of Second Language Writing. 16(3): 194–209.

 

 



[1] Fecha de recepción: 20 de septiembre de 2018 /Fecha de aceptación: 05 de octubre de 2018 /Autor para correspondencia: efernandagonzalez@gmail.com/Universidad Autónoma de Tamaulipas, Unidad Académica Multidisciplinaria de Ciencias, Educación y Humanidades/Dirección: Centro Universitario Victoria, “Lic. Adolfo López Mateos”, Ciudad Victoria Tamaulipas, México, C.P. 871491.