loader image

A Systematic Review of Technologies to Teach Control Structures in Preschool Education

Reading Time: 34 minutes



1. Introduction
In 2006, Jeanette Wing popularized the term “Computational thinking” as a universal set of skills which could allow everyone to use computer science concepts for problem solving (Wing, 2006, 2011). Grover (2018) defined two viewpoints on CT: one is that CT is the cognitive or “thinking” counterpart to practicing computer science in CS classrooms. This means CT is a specific characteristic of practicing computer science and is bound to this discipline. The other viewpoint is that CT is a skill to be integrated by other disciplines and it is a way to approach sense-making in different subjects. Wing’s original definition of CT was broad enough that it ignited educators and policy-makers’ interest in CT (Bocconi et al., 2016). Thus, over the past decade there has been an increase in research around CT interventions targeted at most levels of formal education (Grover and Pea, 2013; Hsu et al., 2018; Yadav et al., 2018; Lyon and Magana, 2020; Stamatios, 2022), its inclusion within other disciplines (Orton et al., 2016; Weintrop et al., 2016; Hickmott et al., 2018), its association with other well-established cognitive skills (Román-González et al., 2017; Robertson et al., 2020; Gerosa et al., 2021; Tsarava et al., 2022), and focusing on creating reliable and valid assessment methods (Tang et al., 2020), amongst others. Moreover, both public and privately-led initiatives have been successfully implemented to foster CT in children and adolescents (Brackmann et al., 2016; Williamson, 2016), as it is regarded as a valuable twenty-first century skill (Yadav et al., 2016).
Several of the most widely accepted and cited definitions of CT propose the use and understanding of control structures such as loops and conditionals as an integral part of CT. For example, Brennan and Resnick (2012) named loops, conditionals and events as central computational concepts in their framework; Grover and Pea (2013) highlighted the use of conditional logic and iteration as well as Shute et al. (2017). In some cases there is no direct reference to control structures in CT definitions, but algorithm design (Khoo, 2020; Saxena et al., 2020) is considered as an essential part of CT. Control structures are basic building components for algorithms (Perkovic, 2015), and therefore an integral part of CT. Moreover, several of the assessments created for evaluating students’ CT in formal education include the evaluation of loops and conditionals, such as Román-González (2015) and collaborators’ CTt; Relkin et al.’s (2020) TechCheck or the CT sections that were incorporated to the PISA mathematics testing in OECD (2019).
Authors such as Bers (2019, 2020) have argued for the inclusion of CT skills in early childhood education, particularly through the use of robots as an embodied, tangible tool which would be intuitive and developmentally appropriate for young children. Teaching young children CT related concepts prepares them to solve real-life challenges in a logical and systematic way, and some authors consider CT as relevant as reading, writing and mathematics (Sanford and Naidu, 2016). The early exposure to computing has potential to engage both boys and girls mitigating gender-related barriers (Manches and Plowman, 2017; Martin et al., 2017).
This work aims at presenting the current state of the art of teaching control structures to preliterate children between 3 and 6 years of age using electronic tools (physical, virtual and hybrid systems) that allow users to construct explicit programs. Our work consists of three parts (see Figure 1): (1) review 1: a systematic literature review (SLR) of reviews aimed at identifying technology used to promote CT in young children; (2) technology overview: an analysis of the characteristics of these tools based on information we found in tool websites and user manuals; (3) review 2: a SLR of empirical evidence related to the use of the tools in teaching control structures to preliterate children between the ages of 3 and 6.

FIGURE 1

Figure 1. An overview of the research pipeline.

The research questions that guide this study are the following:
• What electronic tools exist to support the development of CT in young children? (review 1)
• Which tools are appropriate for preliterate children between the ages of 3 and 6? (technology overview)
• How can children introduce control structures into their programs using electronic tools? (technology overview)
• What tools have been reported to be successful for teaching control structures to young children? (review 2)
In the remainder of the paper we present related works that systematize the knowledge about existing tools that support the development of CT, next we present the methodology and findings of the first SLR that aims to identify existing tools for teaching CT to young children (see Figure 1). In the following step we analyze the tools to identify those that are electronic-based and appropriate for preliterate children between 3 and 6 years old, and provide details related to their price and possibilities that they offer to introduce control structures in children’s code. The resulting list of appropriate tools is used in the second SLR to search for empirical evidence related to teaching control structures to young children. The limitations and results are discussed in the final section of the article and conclusions are laid down.
1.1. Related Work
Previous work has focused on reviewing technological and unplugged tools to promote CT in young children. However, most of the available reviews on this topic focus on the broad aspects of CT and robotics without specifically analyzing the affordances of particular technological tools for learning a specific concept, such as control structures. For example, Silva et al. (2021) focused on describing the available technology for 2–8 year old children as well as curricula implemented for these ages, while Kakavas and Ugolini (2019) focused on they way the teaching of CT has evolved in primary education in the last decades and was successful in identifying the context in which the technology was implemented and in which way CT was assessed. In a recent review (Bakala et al., 2021) we also analyzed the characteristics of robots and activities used in preschool education to promote CT skills with a focus on empirical research, however the specific ways in which each concept encompassed by CT was targeted was not part of our scope. Recent work by Taslibeyaz et al. (2020) shed light into the way studies with young children considered the concept of CT by analyzing its definitions, which skills were targeted and which variables were assessed and included the technological tools used to promote these skills. However, the implications as to how a specific technology causes this improvement and what are the nuances of using different technological tools were not discussed. Similarly, a recent review by Toh et al. (2016) on the use of robots for young children provided context on the type of study conducted and on the effects of robotics on children’s cognitive outcomes as well as parents’, educators’ and children’s opinions regarding the use of these tools. However, the possible benefits are discussed generally regarding robotics and this work does not focus on the outcomes of specific tools. Yu and Roque (2019) provide a comprehensive review of computational toys and kits for young children (7 and under) describing their design features, which computational concepts and practices they target and how they relate to other domains in knowledge. In particular, they analyzed the way conditionals were presented in the technological tools and argued that most of the time conditionals were implemented in an implicit way (thus not represented using explicit if-then statements). In addition, the authors explored the presentation of loops, pointing out many of the available tools used repeat blocks which encapsulated a given sequence, whether digital or concrete. In order to expand upon these findings, this review will focus specifically on the ways technology has implemented control structures and provide an overview of the evidence surrounding these implementations with young children. In this sense, our review will provide a summary of the empirical experiences in which these control structures have been taught to young children and analyze these results. To our knowledge, there isn’t thus far a systematic review of literature which focuses on the implementation of control structures and provides a thorough analysis of how technological tools aimed at early childhood allow its users to learn them. In addition, we conducted a SLR on the existing empirical evidence in which control structures have been taught to children, shedding light into which practices and tools are supported by evidence and thus favorable for its inclusion in the classroom.
2. SLR of Existing Tools (Review 1)
We used a systematic literature review (Kitchenham et al., 2015) to answer our first research question: What tools exist to support the development of CT in young children?
2.1. Methodology
Systematic literature review (SLR) is a method that allows identifying relevant material to a given topic using an objective, analytical, and repeatable approach (Kitchenham et al., 2015). We carried out our literature review following the PRISMA guidelines (Moher et al., 2009). Four reviewers participated in the review process. Firstly, they defined the search term, inclusion and exclusion criteria, and data to extract from the publications. Secondly, two reviewers analyzed the publications to identify the relevant articles. One reviewer extracted the tools from relevant articles. A quality assessment stage was not included, as we were not interested in filtering out low quality studies since we still reviewed each tool or investigating changes in quality over time.
2.1.1. Search Strategy
To identify reviews of technology to support the development of CT in young children we applied an automated search (Kitchenham et al., 2015) in the Scopus search engine (Elsevier Scopus, 2022). The search term was the following:
TITLE-ABS-KEY ( ( ( review AND {computational thinking} AND ( preschool OR child OR {early age} OR kindergarten OR {lower education} OR {early years} OR {elementary education} OR {young learners} OR {primary school} OR {primary education} OR k-6 OR k-8 OR childhood ) ) ) )
We used three keywords: review, computational thinking and childhood (and synonyms) to search in the title, abstract, and keywords.
2.1.2. Study Selection
We defined the following inclusion criteria for the studies’ selection:
• Articles that review electronic-based tools to promote the development of CT in young children.
• Publications focused on children between 3 and 5 years old, including 6 years old, if attending pre-primary school educational level.
Exclusion criteria were:
• Articles written in a language other than English or Spanish.
• Publications that target children older than 6 years.
• Articles limited to unplugged tools.
• Case studies.
• Conference proceedings.
The first round of the selection was made based on the information available in the abstract. Two researchers applied the criteria independently and filter out publications that do not review tools focused on promoting CT in young children. The articles were tagged as “relevant” or “irrelevant.” If an article was classified differently by the reviewers, the full text was reviewed. If there were doubts about an article, they were discussed with two other reviewers that supervised this revision step. Also the articles that were considered relevant by both reviewers were analyzed in detail to confirm or reject their relevance.
2.1.3. Data Extraction
We used a spreadsheet to extract tools found in the publications and articles that mention each tool. We sorted each tool using categories that we developed (see Section 2.4).
2.2. Findings 2.3. Relevant Articles
The search was conducted on 6th of August 2021 and we obtained 54 articles to review (see Figure 2). In the screening phase the reviewers tagged identically 51 of 54 articles reaching an agreement of 0.94%. In the selection process we identified 10 articles relevant for this study. We added to our analysis 3 articles (Kakavas and Ugolini, 2019; Papadakis, 2020; Silva et al., 2021) that were identified by the manual search and that we considered a valuable source of information-Kakavas and Ugolini (2019) that was not indexed by Scopus, Papadakis (2020) that does not contain the word “review” in title, abstract and keywords and Silva et al. (2021) that is a preprint submitted to Elsevier.

FIGURE 2

Figure 2. Steps of the selection process of the first SLR. Reported in line with the PRISMA statement (Moher et al., 2009).

A total of 13 articles (see Table 1) were used to elaborate the list of relevant tools. All the articles were published between 2018 and 2021.

TABLE 1

Table 1. 13 relevant publications that we identified in the first SLR.

2.4. Categories to Classify the Tools
To classify the tools we adapted and expanded categories proposed by Yu and Roque (2019). We obtained 4 main categories (see Figure 3): Physical, Virtual, Hybrid and No information. We divided Physical, Virtual and Hybrid into sub-categories and obtained 9 categories which we used to classify existing tools: Robots with tangible programming interface, Construction kits with no explicit program, Unplugged, Virtual with explicit program, Virtual with no explicit program, Robots with virtual programming interface, Construction kits with virtual programming interface, Virtual tools with tangible programming interface, No information. In the Figure 3 there are more than 8 categories, but only those highlighted in yellow were used to classify the tools.

FIGURE 3

Figure 3. Categories developed to classify the physical aspect of the tools.

We used the category Physical for tools that are fully tangible and do not require screen-based applications. We divided it into Unplugged and Physical tools with electronics. The last category was composed of Robots with tangible programming interface and Construction kits with no explicit program. The category Construction kits with no explicit program contains electronic building blocks that can be connected together to cause certain behavior of the system, but do not require the user to write an explicit program.
Virtual contains fully screen-based tools, such as desktop, mobile, or web apps. This category encompasses tools that do not require the user to write an explicit program (e.g., tools like CompThink App where the user has to solve logical problems without writing code) and those which need an explicit program.
Hybrid tools combine physical and virtual parts. We divided them into Virtual tools with tangible programming interface or Physical tools with virtual programming interface. The first category consists of applications with tangible programming interfaces. The second category is composed of Robots with virtual programming interface and Construction kits with virtual programming interface. The last category contains embedded systems like Arduino that can be programmed using a virtual programming interface.
The “No information” category was used if there was no information about the tool that could be used to classify it.
2.5. Tools
From the 13 relevant publications we extracted 110 tools (106 unique tools). In the case of Code & Go Robot Mouse, we found three different names that referred to this tool: Robot Mouse (Yu and Roque, 2019; Pedersen et al., 2020), Colby robotic mouse (Papadakis, 2020; Bakala et al., 2021) and Code & Go Robot Mouse (Ching et al., 2018; Silva et al., 2021), and we analyzed it as one single tool.
While reviewing the tools mentioned in the articles we found in external sources 4 more tools that we considered relevant for our work. We added Qobo (Physical and Hybrid), VEX 123 (Physical and Hybrid), Sphero indi (Physical and Hybrid), Scottie Go (Virtual) and ended up with a total of 117 tools (110 unique tools, see Table 2).

TABLE 2

Table 2. 117 tools extracted from 13 relevant publications that we identified in the first SLR.

We classified 35 as Physical, 34 as Virtual, 44 as Hybrid and 4 as No information (see Figure 3).
It is important to say that seven tools were present in more than one category (Blue-Bot, Qobo, VEX 123, Sphero indi, VBOT, Makeblock Neuron, Tuk Tuk). For example, Blue-Bot is a robot that can be programmed using buttons on its back and because of that it belongs to the category Robots with tangible programming interface, but there is also a possibility to program it using an application, so it was also classified as a Robot with a virtual programming interface. That is why we refer to 110 unique tools, although we analyzed 117 relevant tools that included duplicated items.
In three cases (Ozobot, LEGO, Robotis and roboplus software) the names that we found in publications were names of brands, not names of specific tools, so it was impossible to classify them, and they were categorized as No information. One publication mentioned Robo Cup Junior as a tool. As far as we know RoboCup Junior (RoboCupJunior, 2022) is an educational initiative, not one particular technology, so we categorized this item as No information as well.
3. Technology Overview
The first aim of this part of our study was to identify how young, preliterate children can introduce conditionals and iterations into their programs using existing tools. This section is motivated by the following research questions:
• Which tools are appropriate for preliterate children between the ages of 3 and 6?
• How can children introduce control structures into their programs using electronic tools?
3.1. Methodology
Four reviewers participated in the revision of existing tools. Two of them reviewed the available online information and extracted the information of interest. The other two participated in the definition of the categories to classify tools’ characteristics and helped to classify doubtful cases.
3.1.1. Tools Selection
We were interested in electronic tools that allow users to construct explicit programs, so we did not further analyze the tools classified as Unplugged, Construction kits with no explicit program, Virtual with no explicit program, and No information.
We identified the relevant tools by filtering out those not appropriate for children between 3 and 6 – tools that target children older than 6 years old or that should be programmed using interfaces that require reading skills (see Table 2). During tool selection we first analyzed the target age of each tool. If the information of the target age was expressed using educational levels like “elementary school” or “kindergarten” we translated this information into age using the United States educational system as reference. If the tool was designed for children older than 6, we tagged it as inappropriate and did not analyze its programming interfaces. If the age was of our interest, we proceeded with the inspection of the user interface. In many cases hybrid tools offered different programming languages/interfaces to cover a wide age spectrum of users, for example, Finch Robot can be programmed using 8 different programming languages and its promotional video states that it is suitable for users from “from kindergarten to college.” In those cases we evaluated only programming languages appropriate for preliterate children. If there was no interface suitable for preschoolers, we marked it as a tool that requires reading skills.
3.1.2. Data Extraction
To collect the information about the tools we reviewed the official websites, video material provided by the manufacturer, online manuals, as well as, youtube videos and amazon websites.
During data extraction we were interested in classifying different types of control structures that can be used with each tool, so we defined categories that we present in Sections 3.2.2.1, 3.2.2.2.
3.2. Findings 3.2.1. Tools Selection
We identified 46 tools (44 unique) appropriate for preliterate children (see Table 3). Twenty Robots with tangible programming interface, 11 Virtual with explicit program and 15 Hybrid tools: 8 Robots with virtual programming interface, 1 Construction kit with virtual programming interface and 6 Virtual tools with tangible programming interface. Two tools (Blue Bot and Sphero indi) were classified as both: Robots with tangible programming interface and Robots with virtual programming interface.

TABLE 3

Table 3. An overview of 46 relevant tools considering their price and possibilities to incorporate control structures into the code.

There were three tools that we analyzed together: KIBO, KIWI and CHERP. KIBO is a robot currently available in the market, formerly known as “KIWI” or Kids Invent with Imagination (Tufts University, 2022). CHERP is a programming language that is used to program KIBO and KIWI, so evaluating CHERP is equivalent to evaluating KIBO and KIWI.
In the case of some tools, the programming interface contained images which made it accessible for preliterate children, but we had the impression that the systems were designed for children older than our target age. They contained text-based challenges (Scottie Go) and menus (BOTS, Neuron App, Move the turtle, RoboZZle), design that we consider unattractive for young children (RoboZZle, BOTS), text-based options with no associated image (“tap” event in Roberto), or comparisons involving high numeric values (Neuron App). Although these tools raised some doubts, we decided to include them in our analysis as we wanted to provide an inclusive overview of the existing tools.
3.2.2. Categories to Classify Control Structures
We developed categories related to the use of control structures to classify tools suitable for young children (see Table 3) that we identified during tools selection step (see Section 3.2.1).
3.2.2.1. Conditionals
To identify how the children can introduce decision making based on certain conditions into their programs we reviewed the programming interfaces and classified the existing tools with categories that we defined in an iterative process. Introducing conditions in the code was typically based on conditional branches (e.g., if-else structures) or based on events (e.g., blocking the program execution until some event occurs). From now on we will refer to those two forms of incorporation of conditions into the code as “conditionals.”
To classify the degree of liberty that the children have while using and building conditionals in their programs, we propose three levels, ordered by increasing complexity for the user:
1. Predefined connection of condition and action: it is possible to use a predefined programming statement that connects an event with an action. For example, the Qobo robot detects coding cards below it and acts according to the statement stored in the card. It has a specific card for conditional turning – if the robot passes over a card with a banana before passing over a bifurcation card, it turns left, but if it passes over a card with an apple, it turns right. Neither the condition nor the resulting action can be modified by the user.
2. Free connection of predefined condition and predefined action: it is possible to combine predefined conditions with predefined actions to build custom conditionals. For example, the Sphero Edu Jr application (see Table 5) allows users to associate a color sensed by the robot (predefined condition) with an action involving movement, light, and/or sound of the Sphero indi robot (predefined actions). The user needs at least two programming statements (condition and action) to build a conditional. In the case of Kodable and RoboZZle these two statements are combined in one coding block: the background color of the block defines the condition (e.g., “if the tile is pink”) and the arrow, the action (e.g., “go right”). The user is able to modify both: the background color and the arrow direction (see Table 5).
3. Free condition building: there are blocks that have to be combined with condition and action. In these cases the user has to use at least three components (bridge-block, condition, and action) to define a conditional. For example, the Matatalab Coding Set contains a “wait until” block that should be combined with a condition (e.g., dark or light) and a sequence of actions in order to build conditionals.
We provide the description and graphical example for each tool that supports conditionals in three tables: Table 4 gathers tools that implement the first level, Table 5 corresponding to the second level, and Table 6 corresponding to the last one.

TABLE 4

Table 4. Tools that allow building conditionals categorized as “Predefined connection of condition and action”.

TABLE 5

Table 5. Tools that allow building conditionals categorized as “Free connection of predefined condition and predefined action”.

TABLE 6

Table 6. Tools that allow building conditionals categorized as “Free condition building”.

The only tools that enable the definition of conditionals using logical operators (e.g., AND, OR) were Makeblock Neuron and Thymio. Neuron online mode allows users to program behaviors using Neuron App, which supports multiple conditions. In the case of Thymio, the user has to associate events sensed by the robot with its behavior. It is possible to combine the sensing and internal state of the robot (e.g., if Thymio touched AND internal state equal to 1) to program advanced robot responses.
In the case of BOTS, Move the turtle, and Makeblock Neuron + Neuron (app) conditionals are based on numerical variables (e.g., a > 5) which makes them more complex than conditionals with non-numerical conditions (e.g., “if the sensed color is red”), as the children have to understand the concept of variable.
In the case of Coding Awbie, the Caution Block is the only means to introduce conditionals into the code, and is a phased out feature as the block is not included in new kits (Getting Started with Osmo Coding Awbie Manual, 2022).
We also analyzed how the code related to a certain condition interacts with the main program, and identified that they occur within either event-based or procedural programming paradigms. Within event-based programming, we identified the following categories:
• Blocking event: the main program contains a condition that blocks the execution until the condition is fulfilled. For example, KIBO contains a “wait for clap” block that makes the robot wait for a clap before executing commands stored in the following blocks.
• Interruption: the main program is interrupted when a certain event occurs. For example, in the case of Pro-bot the main program is interrupted if the sound sensor is triggered and the procedure associated with this event is executed.
• Parallel execution: It is possible for an event to lead to actions to occur in parallel or in addition to those already occurring. For example, an event in Scratch Jr. could generate a sound while a sprite continues moving on the screen.
Using a procedural programming paradigm, we identified the following category:
• Integrated if: the main program contains conditions expressed using the “if” structure that is evaluated during the program’s execution. For example, KIBO allows to incorporate an if-statement into the sequence of commands. If the condition that is evaluated is true, the conditional code is executed and then, the remaining statements.
3.2.2.2. Loops
Another control structure that was relevant for us to analyze was the availability of loops enabling the iteration of commands.
We observed two modalities of implementing the iteration of commands:
• Single command repetition: the tool does not provide the possibility to repeat a sequence of commands, it allows only the repetition of a single action.
• Multiple command repetition: it is possible to repeat multiple commands. In this category we find tools that, due to the design of loop structure, limit the number of pieces that can be repeated (e.g., in Kodable the user is allowed to repeat only two commands) and tools that do not have this restriction.
We also analyzed how the amount of repetitions can be expressed:
• Fixed number of repetitions: the number of repetitions is fixed and cannot be changed by the user.
• Configurable number of repetitions: the amount of repetitions can be defined by the user.
• Infinite loop: it is possible to build infinite loops.
We provide an example for each category in the Table 7.

TABLE 7

Table 7. Examples of tools for categories developed to classify code iteration.

In most cases the amount of repetitions was expressed by associating the number of repetitions with a sequence of statements (similar to a for loop in more advanced programming languages), only BOTS uses exclusively conditions to stop the iteration process (similar to a while loop). KIBO, Finch, Run Marco!, Tynker: Coding for Kids, Scottie Go and VEX 123 offer both types (“repeat X times” and “repeat while”) of repetition statements.
We found many different ways to implement infinite loops: using repeat forever (ScratchJr) or “go to start” command (VEX 123) at the end of the program, elements that contain pieces of code equivalent to “repeat forever” command (Roberto, Code.org), by building circular transitions between states (Dash and Dot), or by calling auxiliary functions (LightbotJr, RoboZZle).
3.3. Cost and Availability
Some tools that we analyzed are currently not available for sale: Plobot is a Kickstarter project that finished in Kickstarter (2022), Robotito, BOTS, Roberto, and T-Maze are academic developments, KIWI is KIBO’s predecessor and is no longer manufactured, Makeblock Neuron and Puzzlets Starter Pack do not appear in online stores and CHERP is a programming language for KIBO and is not sold separately. All these tools were tagged as “unavailable.”
4. SLR of Empirical Evidence (Review 2)
We conducted a second SLR (see Figure 4) to identify literature that reports empirical studies with tools that we considered relevant (see Table 3), in which control structures were taught and/or evaluated in order to respond the following research question: What tools have been reported to be successful for teaching control structures?

FIGURE 4

Figure 4. Steps of the selection process of the second SLR. Reported in line with the PRISMA statement (Moher et al., 2009).

4.1. Methodology
Two reviewers reviewed abstracts and tagged them as “irrelevant” or “relevant.” The last category was used in the cases of publications that meet inclusion criteria or when it was impossible to evaluate the article relevance based on the information available in the abstract. One reviewer reviewed studies that were classified differently among reviewers in the previous step and tried to resolve the doubtful cases. If it was impossible, the articles were considered as “relevant” cases. One reviewer reviewed full-texts of relevant publications and took the final decision about their relevance for this study. We decided not to carry out any quality assessment of the studies as we wanted to provide a broad view of the existing empirical evidence. Two reviewers extracted the data.
4.1.1. Search Strategy
We used an automated search (Kitchenham et al., 2015) in Scopus search engine (Elsevier Scopus, 2022) to identify empirical studies with preschoolers that were developed using tools that we considered relevant (see Table 3). The search term was the following:
TITLE-ABS-KEY ( ( ( {Tool name} AND ( preschool OR child OR {early age} OR kindergarten OR {lower education} OR {early years} OR {elementary education} OR {young learner} OR {primary school} OR {primary education} OR k-6 OR k-8 ) ) ) )
It had two keywords: tool name and young learners (and synonyms) and was used to search in title, abstract and keywords.
In some cases we used curly brackets, that limit the search to exact words, ignoring spelling variation or plurals, around the name of the tool ({Tool name}) to avoid false positive results. For example, in the case of “Coffee Platform” when we used Coffee AND Platform instead of Coffee Platform, the results contained irrelevant publications that did not target the robotic platform. In some cases we excluded publications from areas related to medicine, as some tools’ names were equal to terms used in medicine and also brought false positive results (as in the case of T-Maze). The search term used and the amount of publications found with each tool can be consulted in appendix.
4.1.2. Study Selection
The inclusion criteria for the studies’ selection were the following:
• Articles that report empirical studies with young children using an electronic-based tool that enables activities with control structures.
• Publications that report activities or evaluations focused on control structures.
• Publications focused on children between 3 and 5 years old, including 6 years old, if attending pre-primary school educational level.
Exclusion criteria were:
• Publications that target children older than 6 years.
• Publications that do not report activities or evaluations focused on control structures.
• Off topic articles.
• Articles that describe experiences with users with neurodevelopmental disorders.
• Articles written in a language other than English or Spanish.
• Conference proceedings.
4.1.3. Data Extraction
In the data extraction step we used a spreadsheet to collect information related to the age of participants, number of participants, type of the study, learning outcome, activities aimed at programming conditions, activities that incorporate iterations. Based on the extracted data, two researchers conducted a thematic analysis to summarize study results.
4.2. Findings 4.2.1. Scopus Search Result
The Scopus search for all tools was conducted on 13th of October 2021. In many cases the search brought no results. Only 26 tools of 44 unique tools that we identified, counted with Scopus entries (see Appendix). A total of 205 (202 unique) publications were analyzed. Three publications appeared as repeated because the research that they described involved two relevant tools, for example, Pugnali et al.’s research involved KIBO and ScratchJr, so it was found under the search query for KIBO and ScratchJr. We identified 24 unique publications (see Table 8) that met all inclusion criteria. In the screening phase the reviewers identically tagged 152 of 202 unique articles reaching an agreement of 75%.

TABLE 8

Table 8. 24 relevant publications that we identified in the second SLR.

The 24 relevant publications reported experiences with 10 different tools that we identified as relevant: ScratchJr (evaluated in 7 articles), KIBO (8), KIWI (2), CHERP (2), Code.org (2), BOTS (1), Kodable (1), Move the turtle (1), Strawbies (1) and T-maze (1). Strawbies is an alternative name for Coding Awbie that was used for the search, as the search term with “Coding Awbie” brought no results. Daisy the Dinosaur was mentioned in a study related to Kodable (Pila et al., 2019), but it targets older children (see Table 2). We also found one case of a custom tool (Rose et al., 2017): a game with both ScratchJr-like and Lightbot style programming interface.
4.2.2. Thematic Analysis 4.2.2.1. KIBO/CHERP/KIWI Articles
The only set of technologies for which control structures have been evaluated multiple times was KIBO/CHERP/KIWI, developed by Marina Bers’ group at Tufts University. Of the articles we identified using this technology, five evaluated children’s use of control structures while separating the performance of young children from that of older children, or only including children within our inclusion criteria. All these evaluations used the Solve-It assessments, which were developed by the same research group. Through these assessments, in four of the publications, children who fit our inclusion criteria demonstrated proficiency when programming repeat loops (with a given number of repetitions) and “wait for clap” programs, but were not tested on or were unable to be proficient in the use of sensor loops or conditionals (Strawhacker and Bers, 2015; Elkin et al., 2016; Sullivan and Bers, 2016b; Bers et al., 2019). There was one outlying study where children in Kindergarten were able to demonstrate proficiency across all Solve It assessment areas, including repeat loops, sensor loops, “wait for clap” programs, and conditionals (Sullivan and Bers, 2018). Four other evaluations of this tool did not include specific evaluations of control flow (Kazakoff and Bers, 2014; Sullivan et al., 2017; Bers, 2019; Jurado et al., 2020) while two others did not separate children in our age range of interest from older children.
4.2.2.2. Scratch Jr and Others
Most of the other evaluations involved Scratch Jr. (Papadakis et al., 2016; Portelance et al., 2016; Strawhacker et al., 2018; Pinto and Osório, 2019) and did not evaluate children’s use or understanding of control structures, even though the tool enables the use of control structures. The same happened with evaluations of other systems (Wang et al., 2014; Hu et al., 2015; Rose et al., 2017; Jung et al., 2019; Pila et al., 2019; Arfé et al., 2020; Çiftci and Bildiren, 2020). The evaluations that did include reports on the use of control structures, without an evaluation, involving Scratch Jr., reported either little use or difficulty with control flow blocks (Pugnali et al., 2017; Strawhacker and Bers, 2019). Another included children in our target age, but also older children without separating their performance (Pugnali et al., 2017). One evaluation of LEGO WeDo found some success with repeat loops, but greater success with CHERP (Strawhacker and Bers, 2015).
4.2.2.3. Bottom Line
Only one study (Sullivan and Bers, 2018) provides evidence of children in Kindergarten mastering conditionals and sensor loops. Multiple studies provide evidence of children in our target age group mastering the use of simple repeat loops (repeat a given # of times) or wait for clap programs. The caveat with all these studies is that they are all from the same research group, use the same system, and the same assessment.
With other tools, except for a study of Lego WeDo which also included CHERP (Strawhacker and Bers, 2015), there are no specific assessments of control flow, other than reports of low use or difficulty with using control flow structures for children in our target age range. In other words, in spite of the great diversity of options for children in our target age range to learn about control flow structures, in our review we found only one technology for which there have been multiple empirical studies to understand whether these children can learn how to use these features.
5. Limitations
Although we tried to carry out our study in a systematic way, document all the decisions, and report doubtful cases, the current study still has certain limitations. To complement the tools characteristics related to control structures and cost, we had to appeal to online information. We firstly reviewed official websites and online user manuals, but in some cases the information contained in these sources was not sufficient to answer our research questions. In those cases we reviewed unofficial sources such as youtube videos, blogs and private web pages to complete the missing information. We understand that these are not the most convenient information sources, but we used them if there was no available information through official channels. Another limitation related to our online search is that we reported information that we were able to find, which does not ensure that it is the complete existing information. For example, we reported that the application The Foos allows users to build conditionals of “Free condition building” type based on a youtube video that we found, but we cannot ensure that the tool does not allow building other types of conditionals. There is no free online manual that could provide required information, so to confirm that “Free condition building” is the only type that the tool supports it is necessary to pass all the levels that the game offers, and it was impossible for our team to acquire and personally analyze all the relevant tools. Also, our initial list of tools for young children is limited to the tools reported in scientific publications. It is possible that there are valuable tools that were not mentioned in reviews that we analyzed. We tried to address this issue by adding 3 publications that were not found by SLR and by adding four tools that we found in external sources.
6. Discussion
The present study reviewed the state of the art in the teaching of control structures to young children, specifically preliterate children 3 to 6 years of age. While many of the definitions of CT for young children which gather large amounts of consensus amongst academics describe control structures such as conditionals and loops amongst central aspects of CT (Brennan and Resnick, 2012; Grover and Pea, 2013), how this aspect of CT should be developmentally adapted for young children remains unclear. Our findings suggest there is still a large knowledge gap regarding how children acquire early notions about control structures and what the best tools are to introduce children to these concepts. Despite this, these concepts are often included in the interventions targeted at young children and assessed through specific items in the validated CT tests available for young children (Relkin et al., 2020; Zapata-Cáceres et al., 2020).
Our findings demonstrate that there is a wide variety of technological tools which include robots, virtual applications and hybrids, which aim to teach control structures and are targeted to children of these ages. Thus, we infer it is considered relevant that children acquire these concepts early on. Despite this, our findings regarding the reported classroom based research shows that the specifics of how children learn these concepts through the available tools remains unexplored. None of the systematic review articles we identified presented results that were specific to control structures, instead focusing on broader concepts such as CT (Sullivan et al., 2017), programming literacy (Bers, 2019), or engagement (Pinto and Osório, 2019). Given that CT is an umbrella term which encompasses a wide variety of components such as sequencing, using control structures, abstraction, debugging, amongst others (Shute et al., 2017) we must focus on the specifics of each of them in order to have a better sense of the concept as a whole. This is especially relevant for younger children, as the learning curves for each specific skill might differ with age. So far, we found most of the studies focus on several concepts at once but do not further explore learning outcomes for each activity. Thus, the assessments used were more holistic and successful in detecting general learning and engagement outcomes but lacked information on each of the specific tasks and concepts encompassed. An exception to this general approach was the study reported by Kazakoff and Bers (2014) where they focused specifically on sequencing skills, however we did not find any similar study for the learning of control structures, even though our search targeted this term specifically.
Exploring these aspects is also necessary to determine which approaches provide the adequate affordances to enhance learning of each aspect of CT. For example, in our technology overview we observed several approaches to including the use of control structures in tools, such as interrupting events, active wait, or procedural conditions, however there are currently no studies contrasting the strengths and weaknesses of each of these approaches and whether they produce different results in children’s understanding of the concepts. As a result, there is only evidence of one tool successfully enabling children to learn some aspects of control structures, mainly due to a lack of studies on the use of other tools by young children that include an assessment of control structure use or understanding.
Moreover, future studies on specific tools should focus on the feasibility of their inclusion in the classrooms in a scalable way. Specifically, our findings regarding the cost of several robots suggest some of them are simply too expensive to be available to all children in a given school or classroom. In addition, some of these tools are more adequately design for individual at-home use, which hinders group based-activities thus elevates the cost of its use even more. Thus, so far the use of robots in education at a large-scale would a entail substantial investment for administrators and policy makers, a problem which could be partially subsided through the design of tools with a group-based focus.
The results of our systematic reviews therefore are encouraging in terms of the wide range of approaches designed for young children to learn about control structures, but also identify a large gap in that we know very little about which of these approaches may work better, or how to structure their use. There is therefore a need for future research to further explore the strengths and weaknesses of the available approaches and understand the feasibility of their use in a variety of contexts (e.g., individual vs. shared, home vs. school).
7. Conclusion
The present work demonstrates that there are many diverse tools to support the development of CT in young children. It seems that both academia and industry have interest in designing approaches to enable young children to develop this so-called twenty-first century skill, as we found through our systematic reviews. Although many existing tools allow children to approach advanced programming concepts such as control structures, it is not clear which tools and activities are the most appropriate for teaching them to the youngest programmers. In order to lay the basis for the future research that targets this gap, we provide a systematic overview of existing tools for preliterate children between the ages of 3 and 6. We developed categories that classify the type and complexity of conditionals and iteration structures and used them to categorize each tool. We also provided graphical examples of conditionals that the tools provide.
The analysis of empirical evidence showed that KIBO/CHERP/KIWI is the only tool that consistently demonstrates positive results in teaching control structures to young children. Other tools in our review have not gone through similar evaluations, making it difficult to reach conclusions about their appropriateness for introducing these concepts. The contrast between the diversity of approaches available and the scarcity of evaluations focused on control structures calls for more research, ideally by groups independent of the tools being evaluated, to compare and contrast these approaches in a variety of contexts (e.g., home, preschool).
Data Availability Statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author Contributions
EB, AG, JH, and GTe contributed to conception and design of the study. EB organized the first draft of the manuscript and extracted the tools. EB, AG, and JH wrote sections of the manuscript. EB and AG analyzed the articles involved in the first SLR and reviewed tools characteristics. GTe and JH supervised the revision process. GTr and KP analyzed the articles involved in the second SLR. KP and JH extracted the data and conducted the thematic analysis. All authors contributed to manuscript revision, read, and approved the submitted version.
Funding
This review is a part of a Ph.D. project of the EB that was supported by Comisión Académica de Posgrado (Uruguay).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References

Bakala, E., Gerosa, A., Hourcade, J. P., and Tejera, G. (2021). Preschool children, robots, and computational thinking: a systematic review. Int. J. Child Comput. Interact. 29, 100337. doi: 10.1016/j.ijcci.2021.100337
CrossRef Full Text | Google Scholar

Bers, M. U. (2019). Coding as another language: a pedagogical approach for teaching computer science in early childhood. J. Comput. Educ. 6, 499–528. doi: 10.1007/s40692-019-00147-3
CrossRef Full Text | Google Scholar

Bers, M. U. (2020). Coding as a Playground: Programming and Computational Thinking in the Early Childhood Classroom. New York, NY: Routledge.
Google Scholar

Bers, M. U., González-González, C., and Armas-Torres, M. B. (2019). Coding as a playground: promoting positive learning experiences in childhood classrooms. Comput. Educ. 138, 130–145. doi: 10.1016/j.compedu.2019.04.013
CrossRef Full Text | Google Scholar

Bocconi, S., Chioccariello, A., Dettori, G., Ferrari, A., Engelhardt, K., et al. (2016). Developing computational thinking in compulsory education-implications for policy and practice. Technical report, Joint Research Centre (Seville site).
Google Scholar

Brackmann, C., Barone, D., Casali, A., Boucinha, R., and Muñoz-Hernandez, S. (2016). “Computational thinking: panorama of the americas,” in 2016 International Symposium on Computers in Education (SIIE) (Salamanca: IEEE), 1–6.
Google Scholar

Brennan, K., and Resnick, M. (2012). “New frameworks for studying and assessing the development of computational thinking,” in Proceedings of the 2012 annual meeting of the American Educational Research Association, Vancouver, Canada (AERA 2012) (Vancouver, BC), 1–25.

Ching, Y. H., Hsu, Y. C., and Baldwin, S. (2018). Developing computational thinking with educational technologies for young learners. TechTrends 62, 563–573. doi: 10.1007/s11528-018-0292-7
CrossRef Full Text | Google Scholar

Çiftci, S., and Bildiren, A. (2020). The effect of coding courses on the cognitive abilities and problem-solving skills of preschool children. Comput. Sci. Educ. 30, 3–21. doi: 10.1080/08993408.2019.1696169
CrossRef Full Text | Google Scholar

Elkin, M., Sullivan, A., and Bers, M. U. (2016). Programming with the kibo robotics kit in preschool classrooms. Comput. Sch. 33, 169–186. doi: 10.1080/07380569.2016.1216251
CrossRef Full Text | Google Scholar

Fagerlund, J., Häkkinen, P., Vesisenaho, M., and Viiri, J. (2021). Computational thinking in programming with Scratch in primary schools: a systematic review. Comput. Appl. Eng. Educ. 29, 12–28. doi: 10.1002/cae.22255
CrossRef Full Text | Google Scholar

Gerosa, A., Koleszar, V., Tejera, G., Gómez-Sena, L., and Carboni, A. (2021). Cognitive abilities and computational thinking at age 5: evidence for associations to sequencing and symbolic number comparison. Comput. Educ. Open 2, 100043. doi: 10.1016/j.caeo.2021.100043
CrossRef Full Text | Google Scholar

Grover, S. (2018). A Tale of Two CTS (and a Revised Timeline for Computational Thinking). Blog@ CACM.

Grover, S., and Pea, R. (2013). Computational thinking in k-12: a review of the state of the field. Educ. Res. 42, 38–43. doi: 10.3102/0013189X12463051
CrossRef Full Text | Google Scholar

Hickmott, D., Prieto-Rodriguez, E., and Holmes, K. (2018). A scoping review of studies on computational thinking in k-12 mathematics classrooms. Digit. Exp. Math. Educ. 4, 48–69. doi: 10.1007/s40751-017-0038-8
CrossRef Full Text | Google Scholar

Hsu, T. C., Chang, S. C., and Hung, Y. T. (2018). How to learn and how to teach computational thinking: suggestions based on a review of the literature. Comput. Educ. 126, 296–310. doi: 10.1016/j.compedu.2018.07.004
CrossRef Full Text | Google Scholar

Hu, F., Zekelman, A., Horn, M., and Judd, F. (2015). “Strawbies: explorations in tangible programming,” in Proceedings of the 14th International Conference on Interaction Design and Children (New York, NY), 410–413.
Google Scholar

Ioannou, A., and Makridou, E. (2018). Exploring the potentials of educational robotics in the development of computational thinking: a summary of current research and practical proposal for future work. Educ. Inf. Technol. 23, 2531–2544. doi: 10.1007/s10639-018-9729-z
CrossRef Full Text | Google Scholar

Isnaini, R., Budiyanto, C., and Widiastuti, I. (2019). “Robotics-based learning to support computational thinking skills in early childhood,” in AIP Conference Proceedings (Surakarta).
Google Scholar

Jung, H., Kim, H. J., So, S., Kim, J., and Oh, C. (2019). “Turtletalk: an educational programming game for children with voice user interface,” in Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow), 1–6.
Google Scholar

Jurado, E., Fonseca, D., Coderch, J., and Canaleta, X. (2020). Social steam learning at an early age with robotic platforms: a case study in four schools in spain. Sensors 20, 3698. doi: 10.3390/s20133698
PubMed Abstract | CrossRef Full Text | Google Scholar

Kakavas, P., and Ugolini, F. C. (2019). Computational thinking in primary education: a systematic literature review. Res. Educ. Media 11, 2019. doi: 10.2478/rem-2019-0023
CrossRef Full Text | Google Scholar

Kazakoff, E. R., and Bers, M. U. (2014). Put your robot in, put your robot out: sequencing through programming robots in early childhood. J. Educ. Comput. Res. 50, 553–573. doi: 10.2190/EC.50.4.f
CrossRef Full Text | Google Scholar

Khoo, K. Y. (2020). A case study on how children develop computational thinking collaboratively with robotics toys. Int. J. Educ. Technol. Learn. 9, 39–51. doi: 10.20448/2003.91.39.51
CrossRef Full Text | Google Scholar

Kitchenham, B. A., Budgen, D., and Brereton, P. (2015). Evidence-based software engineering and systematic reviews, volume 4. Boca Raton, FL: CRC Press.
Google Scholar

Lyon, J. A., and Magana, A. J. (2020). Computational thinking in higher education: a review of the literature. Comput. Appl. Eng. Educ. 28, 1174–1189. doi: 10.1002/cae.22295
CrossRef Full Text | Google Scholar

Manches, A., and Plowman, L. (2017). Computing education in children’s early years: a call for debate. Br. J. Educ. Technol. 48, 191–201. doi: 10.1111/bjet.12355
CrossRef Full Text | Google Scholar

Martin, C. K., Pinkard, N., Erete, S., and Sandherr, J. (2017). “Connections at the family level: supporting parents and caring adults to engage youth in learning about computers and technology,” in Moving Students of Color From Consumers to Producers of Technology (IGI Global), 220–244.
Google Scholar

Moher, D., Liberati, A., Tetzlaff, J., Altman, D. G., and the, P. R. I. S. M. A., Group (2009). Reprint–preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. Phys. Ther. 89, 873–880. doi: 10.1093/ptj/89.9.873
CrossRef Full Text | Google Scholar

Orton, K., Weintrop, D., Beheshti, E., Horn, M., Jona, K., and Wilensky, U. (2016). Bringing Computational Thinking Into High School Mathematics and Science Classrooms. Singapore: International Society of the Learning Sciences.
Google Scholar

Papadakis, S. (2020). Robots and robotics kits for early childhood and first school age. Int. J. Interact. Mobile Technol. 14, 34–56. doi: 10.3991/ijim.v14i18.16631
CrossRef Full Text | Google Scholar

Papadakis, S. (2021). The impact of coding apps to support young children in computational thinking and computational fluency. A literature review. Front. Educ. 6, 657895. doi: 10.3389/feduc.2021.657895
CrossRef Full Text

Papadakis, S., Kalogiannakis, M., and Zaranis, N. (2016). Developing fundamental programming concepts and computational thinking with scratchjr in preschool education: a case study. Int. J. Mobile Learn. Organ. 10, 187–202. doi: 10.1504/IJMLO.2016.077867
CrossRef Full Text | Google Scholar

Pedersen, B. K. M. K., Larsen, J. C., and Nielsen, J. (2020). The effect of commercially available educational robotics: a systematic review. Adv. Intell. Syst. Comput. 1023, 14–27. doi: 10.1007/978-3-030-26945-6_2
CrossRef Full Text | Google Scholar

Perkovic, L. (2015). Introduction to Computing Using Python: An Application Development Focus. John Wiley & Sons.
Google Scholar

Pila, S., Aladé, F., Sheehan, K. J., Lauricella, A. R., and Wartella, E. A. (2019). Learning to code via tablet applications: an evaluation of daisy the dinosaur and kodable as learning tools for young children. Comput. Educ. 128, 52–62. doi: 10.1016/j.compedu.2018.09.006
CrossRef Full Text | Google Scholar

Pinto, M. S. M., and Osorio, A. (2019). Aprender a programar en Educacion Infantil: analisis con la escala de participacion. Editorial Universidad de Sevilla 55, 133–156. doi: 10.12795/pixelbit.2019.i55.08
CrossRef Full Text | Google Scholar

Portelance, D. J., Strawhacker, A. L., and Bers, M. U. (2016). Constructing the scratchjr programming language in the early childhood classroom. Int. J. Technol. Design Educ. 26, 489–504. doi: 10.1007/s10798-015-9325-0
CrossRef Full Text | Google Scholar

Pugnali, A., Sullivan, A., and Bers, M. U. (2017). The impact of user interface on young children’s computational thinking. J. Inf. Technol. Educ. 16, 171–193. doi: 10.28945/3768
CrossRef Full Text | Google Scholar

Relkin, E., de Ruiter, L., and Bers, M. U. (2020). TechCheck: development and validation of an unplugged assessment of computational thinking in early childhood education. J. Sci. Educ. Technol. 29, 482–498. doi: 10.1007/s10956-020-09831-x
CrossRef Full Text | Google Scholar

Robertson, J., Gray, S., Martin, T., and Booth, J. (2020). The relationship between executive functions and computational thinking. Int. J. Comput. Sci. Educ. Sch. 3, 35–49. doi: 10.21585/ijcses.v3i4.76
CrossRef Full Text | Google Scholar

Román-González, M. (2015). “Computational thinking test: design guidelines and content validation,” in Proceedings of EDULEARN15 Conference (Barcelona), 2436–2444.

Román-González, M., Pérez-González, J. -C., and Jiménez-Fernández, C. (2017). Which cognitive abilities underlie computational thinking? criterion validity of the computational thinking test. Comput. Hum. Behav. 72, 678–691. doi: 10.1016/j.chb.2016.08.047
CrossRef Full Text | Google Scholar

Rose, S., Habgood, M. J., and Jay, T. (2017). An exploration of the role of visual programming tools in the development of young children’s computational thinking. Electron. J. e-Learn. 15, 297–309. doi: 10.34190/ejel.15.4.2368
CrossRef Full Text | Google Scholar

Sanford, J. F., and Naidu, J. T. (2016). Computational thinking concepts for grade school. Contemporary Issues Educ. Res. 9, 23–32. doi: 10.19030/cier.v9i1.9547
CrossRef Full Text | Google Scholar

Saxena, A., Lo, C. K., Hew, K. F., and Wong, G. K. W. (2020). Designing unplugged and plugged activities to cultivate computational thinking: an exploratory study in early childhood education. Asia-Pacific Educ. Res. 29, 55–66. doi: 10.1007/s40299-019-00478-w
CrossRef Full Text | Google Scholar

Silva, E. F., Dembogurski, B. J., and Semaan, G. S. (2021). A systematic review of computational thinking in early ages. ArXiv, abs/2106.10275. doi: 10.48550/arXiv.2106.10275
CrossRef Full Text | Google Scholar

Stamatios, P. (2022). Can preschoolers learn computational thinking and coding skills with scratchjr? A systematic literature review. Int. J. Educ. Reform. doi: 10.1177/10567879221076077
CrossRef Full Text | Google Scholar

Strawhacker, A., and Bers, M. U. (2015). “I want my robot to look for food”: comparing Kindergartner’s programming comprehension using tangible, graphic, and hybrid user interfaces. Int. J. Technol. Design Educ. 25, 293–319. doi: 10.1007/s10798-014-9287-7
CrossRef Full Text | Google Scholar

Strawhacker, A., and Bers, M. U. (2019). What they learn when they learn coding: investigating cognitive domains and computer programming knowledge in young children. Educ. Technol. Res. Dev. 67, 541–575. doi: 10.1007/s11423-018-9622-x
CrossRef Full Text | Google Scholar

Strawhacker, A., Lee, M., and Bers, M. U. (2018). Teaching tools, teachers’ rules: exploring the impact of teaching styles on young children’s programming knowledge in scratchjr. Int. J. Technol. Design Educ. 28, 347–376. doi: 10.1007/s10798-017-9400-9
CrossRef Full Text | Google Scholar

Sullivan, A., and Bers, M. U. (2016a). Girls, boys, and bots: Gender differences in young children’s performance on robotics and programming tasks. J. Inf. Technol. Educ. 15, 145–165. doi: 10.28945/3547
CrossRef Full Text | Google Scholar

Sullivan, A., and Bers, M. U. (2016b). Robotics in the early childhood classroom: learning outcomes from an 8-week robotics curriculum in pre-kindergarten through second grade. Int. J. Technol. Design Educ. 26, 3–20. doi: 10.1007/s10798-015-9304-5
CrossRef Full Text | Google Scholar

Sullivan, A., and Bers, M. U. (2018). Dancing robots: integrating art, music, and robotics in singapore’s early childhood centers. Int. J. Technol. Design Educ. 28, 325–346. doi: 10.1007/s10798-017-9397-0
CrossRef Full Text | Google Scholar

Sullivan, A., and Bers, M. U. (2019). Investigating the use of robotics to increase girls’ interest in engineering during early elementary school. Int. J. Technol. Design Educ. 29, 1033–1051. doi: 10.1007/s10798-018-9483-y
CrossRef Full Text | Google Scholar

Sullivan, A. A., Bers, M. U., and Mihm, C. (2017). Imagining, Playing, and Coding With Kibo: Using Robotics to Foster Computational Thinking in Young Children. Hong Kong: Siu-cheung KONG The Education University of Hong Kong.
Google Scholar

Tang, X., Yin, Y., Lin, Q., Hadad, R., and Zhai, X. (2020). Assessing computational thinking: a systematic review of empirical studies. Comput. Educ. 148, 103798. doi: 10.1016/j.compedu.2019.103798
PubMed Abstract | CrossRef Full Text | Google Scholar

Taslibeyaz, E., Kursun, E., and Karaman, S. (2020). How to develop computational thinking: a systematic review of empirical studies. Inform. Educ. 19, 701–719. doi: 10.15388/infedu.2020.30
PubMed Abstract | CrossRef Full Text | Google Scholar

Toh, L. P. E., Causo, A., Tzuo, P. W., Chen, I. M., and Yeo, S. H. (2016). A review on the use of robots in education and young children. Educ. Technol. Soc. 19, 148–163.
Google Scholar

Tsarava, K., Moeller, K., Román-González, M., Golle, J., Leifheit, L., Butz, M. V., et al. (2022). A cognitive definition of computational thinking in primary education. Comput. Educ. 179, 104425. doi: 10.1016/j.compedu.2021.104425
CrossRef Full Text | Google Scholar

Umam, M. U. K., Budiyanto, C., and Rahmawati, A. (2019). “Literature review of robotics learning devices to facilitate the development of computational thinking in early childhood,” in AIP Conference Proceedings.
Google Scholar

Weintrop, D., Beheshti, E., Horn, M., Orton, K., Jona, K., Trouille, L., et al. (2016). Defining computational thinking for mathematics and science classrooms. J. Sci. Educ. Technol. 25, 127–147. doi: 10.1007/s10956-015-9581-5
CrossRef Full Text | Google Scholar

Williamson, B. (2016). Political computational thinking: policy networks, digital governance and ‘learning to code’. Crit. Policy Stud. 10, 39–58. doi: 10.1080/19460171.2015.1052003
CrossRef Full Text | Google Scholar

Wing, J. (2011). Research notebook: computational thinking–what and why. Link Mag. 6, 20–23.
Google Scholar

Yadav, A., Hong, H., and Stephenson, C. (2016). Computational thinking for all: pedagogical approaches to embedding 21st century problem solving in k-12 classrooms. TechTrends 60, 565–568. doi: 10.1007/s11528-016-0087-7
CrossRef Full Text | Google Scholar

Yadav, A., Sands, P., Good, J., and Lishinki, A. (2018). “Computer science and computational thinking in the curriculum: research and practice,” in Handbook of Information Technology in Primary and Secondary Education, 1–18.

Yang, K., Liu, X., and Chen, G. (2020). The influence of robots on students’ computational thinking: a literature review. Int. J. Inf. Educ. Technol. 10, 627–631. doi: 10.18178/ijiet.2020.10.8.1435
CrossRef Full Text | Google Scholar

Yu, J., and Roque, R. (2019). A review of computational toys and kits for young children. Int. J. Child Comput. Interact. 21, 17–36. doi: 10.1016/j.ijcci.2019.04.001
CrossRef Full Text | Google Scholar

Zapata-Cáceres, M., Martín-Barroso, E., and Román-González, M. (2020). “Computational thinking test for beginners: design and content validation,” in 2020 IEEE Global Engineering Education Conference (EDUCON) (Porto: IEEE), 1905–1914.
Google Scholar

Appendix

TABLE A1

Table A1. Search term used with each tool to search in SCOPUS.



Source link

share this article
  • This field is for validation purposes and should be left unchanged.

Subscribe to receive the latest business and industry news in your inbox.

  • This field is for validation purposes and should be left unchanged.

latest from the industry
PSYCHEDELICS news

Whitepaper

  • This field is for validation purposes and should be left unchanged.

  • This field is for validation purposes and should be left unchanged.

Use