Research on collaborative learning of computer science has been conducted primarily in programming. This paper extends this area by including short tasks (such as those used in contests like the Bebras Challenge) that cover many other computer science topics. The aim of this research is to explore how problem-solving in pairs differs from individual approaches when tackling contest tasks.
An observational study was conducted on tens of thousands of contestants aged 8–12 years. Statistical tests showed that, compared to individuals, pairs have a higher ratio of correct answers and solve slightly more tasks. They seem to be more successful in some components of computational thinking and are more confident in their answers. In tasks with instant feedback, pairs find the correct solution faster than individuals. As the age of the pupils increases, a trend of decreasing advantages of working in pairs can be observed. These results could be useful for curriculum makers who create computer science textbooks.
This study aims to provide a deeper understanding about the Bebras tasks, which is one of the computational thinking (CT) unplugged activities, in terms of age level, task category, and CT skills. Explanatory sequential mixed method was adopted in the study in order to collect data according to the research questions. The participants of the study were 113,653 school students from different age levels. Anonymous data was collected electronically from the Turkey 2019 Bebras challenge. Factor analysis was employed to reveal the construct validity to determine how accurately the tool measured the abstract psychological characteristics of the participants. In addition, the item discrimination index was calculated to measure how discriminating the items in the challenge were. Qualitative data gathered through the national Bebras workshop was analysed according to content analysis. The findings highlighted some interesting points about the implications of the Bebras Challenge for Turkey, which are discussed in detail. Furthermore, common problems of Bebras tasks are identified and possible suggestions for improvement are listed.
Although Machine Learning (ML) is used already in our daily lives, few are familiar with the technology. This poses new challenges for students to understand ML, its potential, and limitations as well as to empower them to become creators of intelligent solutions. To effectively guide the learning of ML, this article proposes a scoring rubric for the performance-based assessment of the learning of concepts and practices regarding image classification with artificial neural networks in K-12. The assessment is based on the examination of student-created artifacts as a part of open-ended applications on the use stage of the Use-Modify-Create cycle. An initial evaluation of the scoring rubric through an expert panel demonstrates its internal consistency as well as its correctness and relevance. Providing a first step for the assessment of concepts on image recognition, the results may support the progress of learning ML by providing feedback to students and teachers.
As computing has become an integral part of our world, demand for teaching computational thinking in K-12 has increased. One of its basic competences is programming, often taught by learning activities without a predefined solution using block-based visual programming languages. Automatic assessment tools can support teachers with their assessment and grading as well as guide students throughout their learning process. Although being already widely used in higher education, it remains unclear if such approaches exist for K-12 computing education. Thus, in order to obtain an overview, we performed a systematic mapping study. We identified 14 approaches, focusing on the analysis of the code created by the students inferring computational thinking competencies related to algorithms and programming. However, an evident lack of consensus on the assessment criteria and instructional feedback indicates the need for further research to support a wide application of computing education in K-12 schools.
The European Commission Science Hub has been promoting Computational Thinking (CT) as an important 21st century skill or competence. However, "despite the high interest in developing computational thinking among schoolchildren and the large public and private investment in CT initiatives, there are a number of issues and challenges for the integration of CT in the school curricula". On the other hand, the Digital Competence (DC) Framework 2.0 (DigCom) is promoted in the same European Commission Science Hub portal. It shows that both topics have many things in common. Thus, there is the need of research on the relationship between CT and digital competence.
The goal of this paper is to analyse and discuss the relationship between DC and CT, and to help educators as well as educational policy makers to make informed decisions about how CT and DC can be included in their local institutions. We begin by defining DC and CT and then discuss the current state of both phenomena in education in multiple countries in Europe. By analysing official documents, we try to find the underlying commonness in both DC and CT, and discover all possible connections between them. Possible interconnections between the component groups of approaches are presented in Fig.