Argumentation is an important practice in science by which scientists generate, justify and evaluate scientific claims. Argumentation also plays a central role in science education because it leads students toward deep learning by engaging them in the practice of constructing and evaluating scientific arguments. Students therefore need to be explicitly taught and assessed in scientific argumentation (SA) to acquire the competency to engage in the practice. There has, however, been little study on developing assessment instruments for this practice despite the growing interest in the topic. More knowledge is needed about how items can be written and what factors should be considered to promote the demonstration of SA.
This article is based on a three round qualitative study in an iterative process of developing an instrument to assess scientific argumentation competency (SAC) in the context of Physics. In this study, SAC is decomposed into three components with hypothesized increasing cognitive demand: Identification of SA, Evaluation of SA and Production of SA. The focus of this article is to explore how students interact with the assessment instrument and the factors that should be considered to improve the assessment.
In the first round of the study, the author conducted one semi-structured interview with one Chinese Physics teacher, and two think-aloud interviews and two follow-up interviews with students. In the second round of the study, 6 panels including 10 teachers were organized for the instrument review, 4 students were invited for the think-aloud interview, and 30 students attempted the test with 4 students taking part in a follow-up interview. In the third round of the study, 9 students participated in the think-aloud interview while 2 students were invited for the follow-up interview. All the interview data were collected using video/audio calls, and transcribed. The author analyzed all the interview transcripts using thematic analysis.
Five factors emerged from the first round of the study are considered for the revision of the instrument. These are Language, Scenario arrangement, Argument opportunity, Scoring rubric grain size and Construct definition. The five factors were categorized into three aspects, namely, Item design, Scoring rubrics and Construct map according to Wilson (2013)’s “Four Building Blocks” approach. Findings from the first round of the study contribute to the revision of the instrument for the next round. In terms of the second round of the study, both teachers’ and students’ interview generated 4 factors respectively, with 3 factors overlapped. No new factors contributing to the instrument improvement appeared in the third round of the study. In total, there are 10 factors emerging from the study, specifically, Scenario arrangement, Item interrelationship, Explicating the problem to be argued, SA-related terms clarification and expression, Information provided in the task, Understandable and clear language, Scientific accuracy, Test length, Scoring rubric grain size and Construct definition. The author argues that these factors provide detailed information in supporting SAC assessment research although its universality has not been tested.