This study aims to provide a deeper understanding about the Bebras tasks, which is one of the computational thinking (CT) unplugged activities, in terms of age level, task category, and CT skills. Explanatory sequential mixed method was adopted in the study in order to collect data according to the research questions. The participants of the study were 113,653 school students from different age levels. Anonymous data was collected electronically from the Turkey 2019 Bebras challenge. Factor analysis was employed to reveal the construct validity to determine how accurately the tool measured the abstract psychological characteristics of the participants. In addition, the item discrimination index was calculated to measure how discriminating the items in the challenge were. Qualitative data gathered through the national Bebras workshop was analysed according to content analysis. The findings highlighted some interesting points about the implications of the Bebras Challenge for Turkey, which are discussed in detail. Furthermore, common problems of Bebras tasks are identified and possible suggestions for improvement are listed.
Although Machine Learning (ML) is used already in our daily lives, few are familiar with the technology. This poses new challenges for students to understand ML, its potential, and limitations as well as to empower them to become creators of intelligent solutions. To effectively guide the learning of ML, this article proposes a scoring rubric for the performance-based assessment of the learning of concepts and practices regarding image classification with artificial neural networks in K-12. The assessment is based on the examination of student-created artifacts as a part of open-ended applications on the use stage of the Use-Modify-Create cycle. An initial evaluation of the scoring rubric through an expert panel demonstrates its internal consistency as well as its correctness and relevance. Providing a first step for the assessment of concepts on image recognition, the results may support the progress of learning ML by providing feedback to students and teachers.
As computing has become an integral part of our world, demand for teaching computational thinking in K-12 has increased. One of its basic competences is programming, often taught by learning activities without a predefined solution using block-based visual programming languages. Automatic assessment tools can support teachers with their assessment and grading as well as guide students throughout their learning process. Although being already widely used in higher education, it remains unclear if such approaches exist for K-12 computing education. Thus, in order to obtain an overview, we performed a systematic mapping study. We identified 14 approaches, focusing on the analysis of the code created by the students inferring computational thinking competencies related to algorithms and programming. However, an evident lack of consensus on the assessment criteria and instructional feedback indicates the need for further research to support a wide application of computing education in K-12 schools.
The European Commission Science Hub has been promoting Computational Thinking (CT) as an important 21st century skill or competence. However, "despite the high interest in developing computational thinking among schoolchildren and the large public and private investment in CT initiatives, there are a number of issues and challenges for the integration of CT in the school curricula". On the other hand, the Digital Competence (DC) Framework 2.0 (DigCom) is promoted in the same European Commission Science Hub portal. It shows that both topics have many things in common. Thus, there is the need of research on the relationship between CT and digital competence.
The goal of this paper is to analyse and discuss the relationship between DC and CT, and to help educators as well as educational policy makers to make informed decisions about how CT and DC can be included in their local institutions. We begin by defining DC and CT and then discuss the current state of both phenomena in education in multiple countries in Europe. By analysing official documents, we try to find the underlying commonness in both DC and CT, and discover all possible connections between them. Possible interconnections between the component groups of approaches are presented in Fig.
The development of computational thinking is a major topic in K-12 education. Many of these experiences focus on teaching programming using block-based languages. As part of these activities, it is important for students to receive feedback on their assignments. Yet, in practice it may be difficult to provide personalized, objective and consistent feedback. In this context, automatic assessment and grading has become important. While there exist diverse graders for text-based languages, support for block-based programming languages is still scarce. This article presents CodeMaster, a free web application that in a problem-based learning context allows to automatically assess and grade projects programmed with App Inventor and Snap!. It uses a rubric measuring computational thinking based on a static code analysis. Students can use the tool to get feedback to encourage them to improve their programming competencies. It can also be used by teachers for assessing whole classes easing their workload.