The Use of Rubrics in Higher Education and Competencies Evaluation. Scientific Study About Characteristics of Rating Scales Used by Professors
Author(s):
Juan-Carlos Tójar (submitting) Leticia Velasco Martínez (presenting)
Conference:
ECER 2014
Format:
Paper

Session Information

09 SES 04 B, Assessments in Higher Education

Paper Session

Time:
2014-09-03
09:00-10:30
Room:
B012 Anfiteatro
Chair:
Ariane S. Willems

Contribution

Competency-based learning in Spain has achieved a reconfiguration of higher education to meet the current demands of training. In this regard, institutions and different university regulations promote a "constructive alignment" of the essential elements from the curriculum (Biggs, 2005), giving to the competencies the role of curricular model which gives coherence to the new design of degrees.

However, it's enough to review the majority of Educational Programming to detect that it has not been resolved its inclusion in the curriculum (Mateo y Vlachopoulos, 2013) or that, in the most of the case, It only involves comply with the mere process of enunciating a set of competencies (Escudero, 2008), without any strategic-methodological approach in how to integrate the different subjects.

Moreover,  the educational proposals made by the European Higher Education Area (EHEA), such as the use of active methodologies, the self-regulated and independent learning-oriented teaching, the diversification of learning activities (simulations, portfolios, forums), along with the consideration of the multidimensional competencies, require new evaluation tools more dialogic and more compressive than the traditional paper and pencil tests (Ibarrra y Rodríguez-Gómez, 2010; Padilla y Gil Flores, 2008).

In the university context, scoring rubrics are considered an innovation tool to collect evidences of competency acquisition (Baryla, Shelley and Trainor, 2012, White, 2011; Cebrián, 2012; Andradre and Reddy, 2010). Its potential lies in the ability for issuing adjusted valuations about the quality of student’s works in a wide range of subjects or tasks (Blanco, 2008).

In the student's new models competency development, the curriculum isn’t structured as thematic units, but by learning activities (Mateo, 2006). Regarding the activities or tasks evaluated, a question arises: what activities are assessed with assessment rubrics? Are the usual assimilative and reproductive activities, typical of the traditional approach, or more focused tasks to organize and share information, participate in a simulation, self-evaluation a report, etc..?

To get the answers we took as reference the following classification of activities given by Marcelo, Yot, Mayor, Sánchez, Murillo, Rodríguez-Lopez and Pardo (2014): assimilative, information management, application, communication, production, experiential and evaluative.

Moreover, Villa y Poblete (2011, 151), point out that “the difficulty in the evaluation of competencies can be very different depending on the same competencies because some of them are more 'saturated' with knowledge, skills and values than others”. As a result, considering the concept and classification of generic competence by Tuning Project (González and Wagennar, 2003), we propose the following question: What kind of generic competencies are more assessed with rubrics?

Regarding the type of rubric, Blanco (2008, 176) notes that "the selection form one kind or another of rubric depends mainly on the use you want to give to the outcomes, that is, if the emphasis is more focused on the formative aspects or on the summative ones. Other factors to consider: the required time, the nature of the task itself or the specific performance criteria which are being observed”. Based in these assumptions it is studied: What kinds of rubrics are used by teacher: analytic (formative) or holistic (summative)? Do professors know the pedagogical and techniques requirements needed to design rubrics?

These and other issues represent the educational paradigm change with respect to new approaches and tools for evaluating competencies. In this theoretical framework it is set up a research in order to know the aims pursued by professors when designing an evaluation rubric. In addition, we analyze the types of rubrics that professors use to support and guide the teaching and learning processes.

Method

Two hundred rubrics from six different Spanish public universities were analyzed. Sampling was intentional, considering the public access to them and compliance with some minimum requirements: identification data and basic elements of their structure according to the specialized literature This research combines quantitative and qualitative methodology. The quantitative perspective is part of ex post facto correlational studies, and qualitative perspective was established through inductive-deductive procedures of categorization from the units of analysis (the rubrics). The category system was developed through a qualitative content analysis of the rubrics, as proposed by Miles and Huberman (1985, cited in Tójar, 2006). These authors consider the qualitative analysis as a set of tasks that interact itself. a) Reduction of information. It was made before data collection in order to define the conceptual framework of the study, preliminary purposes, the first hypotheses and the implicit questions in the theoretical foundations. Further analysis procedures and data processing were implemented. This phase is composed of a series of processes that also interact between itself:  The separation in units was made through the analysis of information (rubrics), considering the different thematic criteria which emerged from the authors proposal's who explained the uses and technical requirements of this tool.  The identification and classification of units was made using a mixed model of categorization (Pinto and Grawitz, 1967, cited in Tójar, 2006, 476) to develop predefined categories (deductive), derived from the specialized literature on rubrics, and ad hoc categories (inductive), which were constructed from the observation itself. Tree diagrams for the categorization were used.  The synthesis and grouping of the units was produced from the mixed categorization process (deductive-inductive) to construct a useful system of categories for collecting data contained in rubrics. Macrocategory 1. Purposes of evaluation rubrics: quizzes, activities and generic competencies. Macrocategory 2. Branches of knowledge. Macrocategory 3. Kinds of rubrics: introduction, kinds of analytical rubrics, explanation, coherence, homogeneity, continuity, parallelism, amplitude, levels of execution, tag and scales. a) The organization and transformation of data. In this phase, the information was specified and organized with descriptive table. b) The extraction and verification of conclusions allowed the identification of regularities and patterns, resulting in certain generalizations, typologies and models.

Expected Outcomes

In relation to the purposes of the professor in the design of rubrics, the tasks more evaluated are written documents about the acquisition of skills in a traditional way (e.g. Written reports). In contrast, simulation tests are less evaluated because they involved higher integration and change of cognitive resources on the real contexts (e.g. “internships”). This type of evidence is more consistent with competency-based training that aims to integrate into the labor market. Regarding the types of evaluated activities, the productive nature is best evaluated in than the experiential one. A possible lack of applicability and transferability principles promoting EHEA stands. With regard to the generic competencies, the results warn that professors can be evaluating only instrumental competencies, or in conjunction with the systematic ones, omitting interpersonal competencies considered of equal or greater importance in the academic and professional training. In addition, technical and / or pedagogical criteria that consider professors when designing rubrics are: parallelism in language, an adequate number of levels of performance or consistency, uniformity and continuity of the evaluation criteria. By contrast, the rubrics that have more ambiguity in their interpretation are the specific ones, with a disciplinary character, with punitive tags, which contain a minimum range in the description of the competence or lacking an explanation of the criteria evaluation. Por último, respecto a la relación entre categorías, las rúbricas analíticas específicas son más utilizadas para evaluar documentos escritos debido a que solo evalúa la adquisición de conocimientos disciplinares o habilidades instrumentales. Finally, regarding the relationship between categories, the specific analytic rubrics are more used to assess written documents because they only evaluate the acquisition of disciplinary knowledge and instrumental competencies.

References

Baryla, E., Shelley, G. & Trainor, W. (2012). Transforming Rubrics Using Factor Analysis. Practical Assessment, Research & Evaluation, 17 (4). Biggs, J. B. (2005). Calidad del aprendizaje universitario. Madrid: Narcea. Blanco, A. (2008). Las rúbricas: un instrumento útil para la evaluación de competencias, en L. Prieto (Ed.). La enseñanza universitaria centrada en el aprendizaje. (pp. 171-188). Barcelona: Octaedro. Blanco, A. (2011). Tendencias actuales de la investigación educativa sobre las rúbricas. En K. Bujan, I. Rekelde, & P. Aramendi, La evaluación de competencias en la Educación Superior. Las rúbricas como instrumento de evaluación. (pp. 59-74). Seville: MAD. Cebrián, M. (2012). (Ed.). Presentación. Congreso Internacional de Evaluación mediante e- Rúbrica. Málaga: University of Malaga. Escudero, J. M. (2008). Las competencias profesionales y la formación universitaria: posibilidades y riesgos. REDU, 1. González, J, & Wagenaar, R. (2003). Tuning Educational Structures in Europe. Informe final Fase 1. Bilbao: Universidad de Deusto. Ibarra M.S & Rodríguez-Gómez, G. (2010). Aproximación al discurso dominante sobre la evaluación del aprendizaje en la Universidad. Revista de Educación, 351, 385-407. Marcelo, C., Yot, C., Mayor, C., Sánchez, M. Murillo, P., Rodríguez-López, J. M. & Pardo, A. (2014). Las actividades de aprendizaje en la enseñanza universitaria: ¿hacia un aprendizaje autónomo de los alumnos? Revista de Educación, 363, 334-359. Mateo, J. (2006). Claves para el diseño de un nuevo marco conceptual para la medición y evaluación educativas. Revista de Investigación Educativa, 24 (1), 165-189. Mateo, J. & Vlachopoulos, D. (2013). Evaluación en la universidad en el contexto de un nuevo paradigma para la educación superior. Educación XXI: Revista de la Facultad de Educación, 16 (2), 183-207. Padilla, M. T. & Gil, J. (2008). La evaluación orientada al aprendizaje en la Educación Superior: condiciones y estrategias para su aplicación en la docencia universitaria. Revista Española de Pedagogía, 241, 467-486. Santos, M. A. (1999). 20 paradojas de la evaluación del alumnado en la Universidad española. Revista Electrónica Interuniversitaria de Formación del Profesorado, 2 (1). Reddy, Y. M. & Andrade, H. (2010). A review of rubric use in higher education. Assessment & Evaluation in Higher Education, 35 (4), 435-448. Tójar, J. C. (2006). Investigación cualitativa. Comprender y actuar. Madrid: La muralla. Villa, A. & Poblete, M. (2011). Evaluación de competencias genéricas: principios, oportunidades y limitaciones. Bordón, 63 (1), 147-170.

Author Information

Juan-Carlos Tójar (submitting)
University of Malaga, Spain
University of Málaga
Málaga

Update Modus of this Database

The current conference programme can be browsed in the conference management system (conftool) and, closer to the conference, in the conference app.
This database will be updated with the conference data after ECER. 

Search the ECER Programme

  • Search for keywords and phrases in "Text Search"
  • Restrict in which part of the abstracts to search in "Where to search"
  • Search for authors and in the respective field.
  • For planning your conference attendance, please use the conference app, which will be issued some weeks before the conference and the conference agenda provided in conftool.
  • If you are a session chair, best look up your chairing duties in the conference system (Conftool) or the app.