Cs First Game Design Sprite Are Disappearing
Introduction
This article describes the design and evaluation of a novel educational programming game, Pirate Plunder, to teach procedural abstraction to children. Pirate Plunder aims to teach children to identify code duplication in Scratch projects and be able to remove this using procedural abstraction and code reuse. Despite these skills being an important part of computer science, children, particularly those in primary education (age 5 to 11), are rarely taught them because of a lack of teacher expertise (Rich et al., 2019) and school resources (Larke, 2019). The article builds upon a previous report on the design of Pirate Plunder (Rose et al., 2018) by describing in more detail how the game introduces the learning content. It provides a complete and final report of an extended evaluation described in Rose et al. (2019). As such, it includes a detailed description of the complete study (employing a crossover design) and a comprehensive discussion of the complete set of results and their implications.
RQ1 – Can a game-based learning approach be used to teach primary school children to use procedural abstraction in Scratch projects?
RQ2 – What aspects of the game design influence the effectiveness of this approach and why?
This section gives an overview of the context behind these research questions, including programming tools in computer science education, code smells and procedural abstraction, before discussing similar work and how Pirate Plunder has been designed to introduce these concepts.
Computer Science Education
Over the last few years, ICT-focused computing curricula in compulsory education have been replaced with wider topics of informatics, digital literacy and computer science (Heintz et al., 2016). This shift has aimed to give children the skills and knowledge required in an increasingly digital world, with learning content that focuses on computer programming, robotics and computational thinking. This new type of computing curriculum was introduced in England in 2013, in part due to recommendations by Livingstone and Hope (2011) and The Royal Society (2012) in reports that were critical of the old curriculum.
Computing education in English schools is still described as "patchy and fragile" (Furber, 2012, p. 6), with teachers "lacking sufficient theoretical and technical knowledge of computing" (p. 54). This stems from the curriculum being introduced without adequate training programs, infrastructure, and materials. This problem is particularly pronounced in primary education where teachers are not subject specialised. It is often left to teachers to develop an understanding of the required learning content by themselves (Rich et al., 2019), meaning that computing is often left out of classroom teaching entirely (Larke, 2019) or children are given educational programming tools with little or no support.
Educational Programming Tools
There is a multitude of educational programming tools available for use in computer science education (e.g. Cooper et al., 2003; Hooshyar et al., 2016; Weintrop & Wilensky, 2013). These differ in type, cost, complexity and learning approach, meaning that teachers are often unsure which they should be using. The problem is exacerbated by the limitations in hardware and financial restraints in schools. Yet, there is definite potential for these tools to be used to support children's understanding of key concepts in computing.
These tools include visual programming environments, games, physical devices and unplugged activities. They range from symbolic drag-and-drop programming to text-based programming languages that allow procedures, variables, iteration and conditional execution (Duncan et al., 2014). Novices can find it easier to learn computer science using block-based languages, relative to text-based languages, because they rely on recognition instead of recall (blocks are selected from a pallet), reduce cognitive load by chunking code into smaller numbers of meaningful elements and allow users to avoid basic errors by providing constrained direct manipulation of structure (Bau et al., 2017).
The most widely used educational programming tool is Scratch, with over 53 million projects shared on its online platform since its public release in 2007 (Scratch Team, 2020). It is also the most popular environment in primary education (Rich et al., 2019). Scratch is a visual programming environment designed for children age 8 and above. It aims to "introduce programming to those with no previous programming experience" (Maloney et al., 2010, p. 2). Scratch uses a block-based programming language in which blocks are combined to form scripts. Its design is inspired by constructionism (Papert, 1980), a learning theory where knowledge and problem-solving skills are developed through exploration. Scratch supports constructionism by always having the block-palette visible, having little in-built guidance and feedback and no error messages. This type of block-based programming is also prevalent in other popular tools used in primary education including Code.org (Code.org, 2020), Tynker (Neuron Fuel, 2020), Hopscotch (Hopscotch Technologies, 2020) and Purple Mash (2Simple Ltd, 2020).
Computational Thinking
It is difficult to discuss computer science education without acknowledging the idea of computational thinking and its implications. Computational thinking stems from the idea that computer science can help develop wider problem-solving and logical thinking skills. It involves skills such as working at multiple levels of abstraction, writing algorithms, understanding flow control, recognising patterns and decomposing problems (Rose et al., 2017). Proponents of computational thinking take this one step further, suggesting that these skills are a "foundational competency for every child" (Grover et al., 2018, p. 1). As such, computational thinking has been used by policymakers as justification for introducing computer science into compulsory education. Yet, it has been criticised by some for its 'decoupling' from the theoretical foundations of computer science, along with the lack of evidence for it as a multidisciplinary problem-solving skill (Denning, 2017).
Nonetheless, in this article, we concentrate on the use of these skills in computer programming, using the computational thinking measures Dr. Scratch (Moreno-León & Robles, 2015), to measure abstraction in Scratch projects, and the Computational Thinking test (Román-González et al., 2018), used as a secondary measure of improvements on the computational thinking skills used in programming.
Code Smells and Bad Programming Practices
The constructionist programming approach in block-based environments like Scratch can result in poor programming practices because proper software engineering principles are not introduced (Dorling & White, 2015). Code smells are a useful method for identifying these bad practices. The term 'code smell' was coined by Fowler (1999) and refers to a surface indication in a program that usually corresponds to a deeper problem. Identifying code smells and 'refactoring' code to remove them can help improve the design and readability of code. Refactoring is the "process of changing a software system in such a way that it does not alter the external behaviour of the code but improves its internal structure" (p. 9). Fowler gives a list of possible code smells found in object-oriented programming, including duplicated code, long methods, large classes, and long parameter lists.
Code Smells in Scratch
Similar code smells are also found in block-based programming languages like Scratch. As stated in "Educational Programming Tools" section, Scratch supports a constructionist programming approach, where solutions are unplanned and created largely through self-directed exploration or 'tinkering' (Maloney et al., 2010). This allows learners to quickly create programs. However, it can result in the learner producing 'bad' code (indicated by code smells) and forming bad programming habits (Meerbaum-Salant et al., 2011). This is particularly significant because of the widespread use of Scratch in compulsory education and the lack of relevant teacher knowledge within the profession.
The most common code smells in Scratch projects are duplicated blocks (a repeated sequence of blocks, regardless of block inputs, that are used to reuse code), large scripts (a long script that could be shortened through better code reuse) and dead code (blocks that can be safely removed from the program without affecting its behaviour). Table 1 shows the code smell prevalence in four exploratory analyses of large Scratch project repositories.
| Table 1. Scratch Project Analyses for Code Smells. |
Hermans and Aivaloglou (2016) found that code smells can impact understanding, debugging and the ease with which learners can alter projects. Furthermore, Techapalokul and Tilevich (2015) found that novice programmers are "prone to introducing some smells continue to do so even as they gain experience" (p. 10). Code smells are particularly important because 'remixing' other users' projects is a large part of the Scratch online platform (Dasgupta et al., 2016).
Procedural Abstraction and the Extract Method
Procedural abstraction can be used to remove code smells. It is an important computer science skill and is one of the two kinds of abstraction utilised in computer science, with the other being data abstraction (Haberman, 2004). Procedural abstraction involves the separation of the logical properties of an action from the implementation details. This is often done by moving fragments of code into a procedure (with or without data arguments to pass information to it) that can then be invoked in multiple places, a process known as the 'extract method' (Fowler, 1999). Kallia and Sentance (2017) describe procedural abstraction (and its subconcepts) as a potential 'threshold concept' in computer science. That is, a concept that opens "up a new and previously inaccessible way of thinking about something" (Meyer & Land, 2003, p. 1). However, procedural abstraction is difficult for novices and an area where misconceptions often arise.
Duplicated blocks, large scripts and dead code smells in Scratch can all be removed using procedural abstraction and the extract method. This is done using 'custom blocks' (procedures in Scratch) that the user can define. These can be given 'inputs' (parameters) to pass data to the block. For example, Figure 1 shows an example of a Scratch project that contains a duplicated code smell: the 'point in direction', 'repeat' and 'move' blocks are all repeated four times with different input values. Figure 2 shows the same project with the duplicated blocks 'extracted' into a custom block called 'turnAndMove' that takes two arguments, distance and degrees. These projects are shown using Scratch 2, as Scratch 3 had not been released when this study was conducted. Scratch is limited in that custom blocks can only be used within the sprite they are defined, a limitation that has led to calls for a functionality change (Techapalokul & Tilevich, 2019b).
Figure 1. Moving a Cat Sprite Around the Four Corners of a Park Using Block Duplication.
Figure 2. Using the Extract Method and a Custom Block to Refactor the Project in Figure 1 .
In their Scratch project analysis, Robles et al. (2017) found that the use of custom blocks and cloning (duplicating sprites at runtime) did not impact the amount of code duplication. This supports the idea that procedural abstraction is difficult for novices, particularly children, to use appropriately.
Game Design to Introduce and Reinforce Procedural Abstraction
So far, we have highlighted that children are being taught computer science in primary education using block-based programming tools such as Scratch, with teaching that is often limited by school resources and teacher expertise. The open-ended design of these block-based tools can allow users to produce 'bad' code, which can be recognised through code smells, such as duplicated blocks and long scripts. These code smells can be refactored through the extract method, a skill that requires procedural abstraction, itself an important part of computer science.
We hypothesise that game-based learning can play a key role in supporting computer science learning in block-based visual programming environments. Procedural abstraction is a good focus for this because it is challenging for primary school children and is not part of the current English national curriculum.
Related Work
There have been several efforts to teach the higher-level principles of abstraction to children, getting them to think about and articulate problems at different levels of abstraction (e.g. Rijke et al., 2018; Statter & Armoni, 2020). However, research on procedural abstraction in block-based programming tools is limited. Sherman and Martin (2015) use it as part of their computational thinking rubric in the App Inventor block-based tool (Wolber, 2011). Techapalokul and Tilevich(2019a) created an automated refactoring tool for Scratch that removes code smells using the extract method. Although similar, our work differs from this in that we aim to teach children to do this process themselves and to understand why this is beneficial. Kalas and Benton (2017) explored the factors that underpin primary school children's understanding of procedural abstraction and suggest a pedagogical approach that they found effective. Pirate Plunder aims to do this without teacher instruction, which leads us to our first research question (RQ1): can a game-based learning approach be used to teach primary school children to use procedural abstraction in Scratch projects?
Procedural abstraction is introduced in several existing educational games. How this is implemented depends on the type of programming interface used, which can be split into two categories: games that use symbolic programming interfaces with minimal text, including Lightbot (Yaroslavaski, 2014) and AutoThinking (Hooshyar et al., 2019), and those that use text-heavy block-based languages like Google Blockly (Google, 2020), including ctGameStudio (Werneburg et al., 2016) and Dragon Architect (Bauer et al., 2017). In the first category, procedural abstraction is implemented similarly to Lightbot, which gives the player one or two procedures that can be called from the main set of instructions (Figure 3 ). AutoThinking uses a similar approach, allowing the player to save sets of instructions to procedures that can then be executed using their number identifier from the main program. Whilst these are useful ways of introducing procedural abstraction, they do not allow players to use parameters to pass data to their procedures. Understanding parameters is an important part of learning to use procedures as it allows for further code reuse. In the second category, procedures are created and used from the block pallet similar to Scratch ("Procedural Abstraction and the Extract Method" section). Pirate Plunder falls into this category but aims to teach procedural abstraction using a block-based syntax that is similar to Scratch, which is simpler than Google Blockly. The rest of this section explains the novel approach to introducing this content: concept scaffolding, tutorials and feedback and customisation, which leads into RQ2: what aspects of the game design influence the effectiveness of this approach and why?
Figure 3. How Lightbot Introduces Procedural Abstraction.
Concept Scaffolding
Pirate Plunder introduces procedural abstraction through a series of scaffolded concepts that are designed to justify the rationale behind using them. This focuses on both applied knowledge and contextualised theoretical knowledge, in line with recommendations by Kazimoglu et al. (2011) when introducing computer programming concepts in serious games. The game begins with basic Scratch blocks (events and motion) that do not require any prior Scratch knowledge. The learning content is introduced intrinsically as part of the game, an approach that Habgood and Ainsworth (2011) found to be effective in educational games.
We designed a learning trajectory for the game to scaffold the concepts in a way that children would understand (Figure 4 ). Learning trajectories are used in mathematics to define the developmental progression to a mathematical goal, giving a set of instructional tasks that help children develop higher levels of thinking (Clements & Sarama, 2004). The goal of Pirate Plunder is to introduce procedural abstraction. The levels are designed as instructional tasks to help the player understand the concepts along the trajectory, with the Scratch blocks or functionality for that concept only made available to the player once they have reached a certain level.
Figure 4. Pirate Plunder Learning Trajectory.
The difficulty progression is described in more detail in previous work (Rose, 2019). In summary, the levels in each stage (or concept) get progressively more difficult, requiring more duplication and longer scripts. For each subsequent stage, the player is then introduced to a technique or block that can be used to remove that duplication, moving along the learning trajectory. For example, in the final 'loops' level, the player must use three duplicated 'repeat 5, move 1' blocks and two duplicated 'repeat 2, move 1' blocks (Figure 5 ). Figures 6 and 7 show how this could be refactored using custom blocks (procedures) and custom blocks with two inputs (a parameterised procedure). Concept scaffolding in Pirate Plunder is particularly important because it progresses from procedures to parameterised procedures, which are one of the most challenging concepts in introductory programming (Madison & Gifford, 1997).
Figure 5. The Final 'Loops' Level Before Custom Blocks Are Introduced.
Figure 6. The Level in Figure 5 Completed Using Two Procedures (No Parameters).
Figure 7. The Level in Figure 5 Completed Using a Parameterised Procedure.
The final concept in the learning trajectory is instances of sprites. In Scratch, this is done through cloning, which allows a sprite to create a clone of itself or another sprite whilst the project is running. This duplicate is a separate instance of the original or parent sprite but will inherit properties from the parent that can be modified. Cloning a sprite is similar to creating an instance of a class in object-oriented programming, where an 'instance' is a concrete occurrence of an object that is created during runtime. Although not part of procedural abstraction, it was included in the learning trajectory because Dr. Scratch uses it as a measure of abstraction (Moreno-León & Robles, 2015) and this was our primary measure during the study.
Pirate Plunder contains 40 levels that were used in the data analysis (Figure 8 ). There are also eight 'general' levels designed as a further challenge to players who have completed the game. These all unlock when the player completes level 40.
Figure 8. Pirate Plunder Player Avatar and Level Select (General Levels Are Indicated With Black Bars).
The game motivates players to use the taught functionality through block limits, collectable items, required block validation and obstacles. Each challenge limits the number of total blocks that can be used in the program, forcing the player to address block duplication and produce an optimal solution.
Tutorials and Feedback
Pirate Plunder combines 'process constraints': increasing the number of features (in this case, blocks) that the player can control as they progress through the game, with 'explanations' that specify exactly how to perform an action (Lazonder & Harmsen, 2016). These explanations are made in 'tutorial' levels (Figure 9 ) that introduce the player to the next concept on the learning trajectory. The player then uses these concepts or blocks in 'challenge' levels.
Figure 9. A Stage of the 'Custom Blocks' Tutorial Level in Pirate Plunder, in Which the Player Is Guided Through the Tutorial by the Red Parrot in the Top Left Corner and Is Given Feedback on Each Stage by the Green Parrot in the Top Right.
Tutorial levels guide the player through using each concept or block. This approach was inspired by the moving tutorial character from Dragon Architect (Bauer et al., 2017), a block-based programming game, and Stagecast Creator (Seals et al., 2002), a rule-based visual programming environment with a section that walks the player through the application functionality. Pirate Plunder uses a combination of these two approaches. Players must follow a series of actions within a tutorial that are highlighted on-screen by a moving help character (the red parrot). They must follow the actions correctly to complete the level. These differ from 'challenge' levels, where the player is not guided through the level by the help character. The combination of tutorials, challenges and the learning trajectory was designed as a novel way to deliver conceptually difficult content. Later tutorials show refactored challenge levels that the player has already seen, demonstrating the justification for using those blocks later in the game.
Players are given feedback on both tutorial and challenge levels by the green parrot (Figure 10 ). This is automatically given for guidance, warnings and level validation. Guidance feedback is for general block use and program issues, such as missing an 'event' block to start program execution. Warnings are for behaviour that might break the game (e.g. recursion). Level requirement feedback is given for not using the required blocks and reaching the block limit.
Figure 10. A Pirate Plunder Challenge Level, in Which the Green Parrot Gives Both Hints (When Clicked on) and Feedback.
On challenge levels, the player can also ask the green parrot for hints on how to complete the level. The player is given between three and six hints designed to guide the player. Table 2 shows the hints given on the first procedures challenge.
| Table 2. The Hints Given on the First Pirate Plunder Procedures (Custom Blocks) Challenge. |
Customisation
Customisable player avatars are an important part of keeping players motivated throughout Pirate Plunder. Each player has an avatar that they can purchase items for using coins collected when playing through levels Figure 8 . These items unlock as the player progresses through the game. Bailey et al. (2009) showed that there is a strong link between self-designed avatars and game enjoyment, as players identify with and become invested in their character. As an additional motivator, the Pirate Plunder login screen has players select their avatar from a list of all the avatars in their class.
Method
Participants
The participants were 91 children between 10 and 11 years old (M = 10.58, SD = 0.32) from a large primary school in northern England. They were largely inexperienced with Scratch (having had sporadic lessons throughout primary school) and had no experience with the custom blocks or cloning that were used as a measure of abstraction in this study.
Experimental Design
The study followed a pre-to-post-test, partial-crossover, quasi-experimental design to measure for improvements using procedural abstraction in Scratch after playing Pirate Plunder (Figure 11 ). For the first phase of the study, the three groups were split into Pirate Plunder (intervention), spreadsheets (non-programming active control) and Scratch (programming active control). In phase 2, the two control groups then crossed over to Pirate Plunder and the intervention group to the spreadsheets curriculum. The crossover was done for ethical reasons (so that all participants had the opportunity to play the game) and for game analytics (with three times as many players playing the game, more concrete conclusions can be drawn from the data to address RQ2). It also enabled us to see whether the phase 1 control groups would experience the same improvement on the assessments after playing the game.
Figure 11. Diagram of the Experimental Design.
Participants were assessed for their Scratch baseline ability using a Scratch 'baseline' task, their ability to use procedural abstraction in Scratch through a different Scratch 'challenge' and a multiple-choice Scratch abstraction test, and their computational thinking ability using the Computational Thinking test (Román-González et al., 2017). After playing Pirate Plunder, all participants were given questionnaires to measure their confidence using Scratch and 45 of them took part in artifact-based interviews.
Intervention Materials
Pirate Plunder
Pirate Plunder was used for the intervention condition of the study. Participants played through the game at their own pace, with one researcher available to support those that had any difficulty. They continued each session where they left off in the previous session. Player performance is reported in "Pirate Plunder Player Performance" section.
Spreadsheets Curriculum
The main comparison in the study was between Pirate Plunder and a spreadsheets curriculum designed for the age group produced by the UK-based educational resources company, Twinkl Educational Publishing(2018). It was chosen as the primary control group activity because it did not involve programming (beyond using pre-made formula to calculate values, e.g. SUM and AVERAGE) or explicit computational thinking, yet still had participants using technology and being exposed to new learning content (Table 3).
| Table 3. Spreadsheets Curriculum Lesson Breakdown. |
Scratch Curriculum
A Scratch curriculum (also produced by Twinkl) was used as a secondary comparison during phase 1. It involved creating an animated story based on a 'haunted house' and did not cover custom blocks or cloning (Table 4).
| Table 4. Scratch Animated Stories Curriculum Lesson Breakdown. |
Both curricula were delivered by the first author and contained six-hours of learning content, the same amount of time that participants were given to play Pirate Plunder during the intervention. The possibility of potential experimenter bias is discussed in "Potential Experimenter Bias" section.
These two control groups were used to represent usual practice for computing lessons in England, representing a tougher test than usual (non-computing) lessons. They were not taught procedural abstraction as it is not part of the curriculum for this age group, with the aim of the study being whether they could learn it through a game-based approach. Comparing Pirate Plunder with a procedural abstraction curriculum delivered using traditional teaching is the aim of a future study ("Future Work" section).
Instruments and Measures
Scratch Baseline Task
The Scratch baseline task was designed by the first author to allow participants to demonstrate Scratch proficiency, but to a specification that involved duplication, enabling them to use procedural abstraction if they were able to. The task involved animating the cat sprite around the edges of a rectangle, leaving an object on each corner. Figure 12 shows the project that participants started with. The Scratch baseline task was used at pre-test instead of the Scratch challenge assessment because it enables participants to achieve an outcome without prior Scratch knowledge. Whereas the Scratch challenge ("Scratch Challenge" section) requires specific functionality that the participants had not been taught before the study.
Figure 12. Scratch Baseline Task Starter Project.
Both the Scratch baseline task and Scratch challenge were assessed using the Dr. Scratch automated assessment tool (Moreno-León & Robles, 2015), as it gives a quantitative score for 'abstraction and decomposition' (Table 5).
| Table 5. Dr. Scratch Abstraction and Decomposition Scoring System. |
Scratch Challenge
The Scratch challenge was designed by the authors to see if participants could use procedural abstraction to reduce the block count in a pre-made project that contained both duplicated blocks and sprites (Figure 13 ). The project animates a cat sprite around a map, leaving a lamp post sprite on each corner. The optimal solution uses custom blocks and cloning to reduce the number of blocks and sprites but still achieve the same functionality (Figure 14 ). As with the Scratch baseline task, the Scratch challenge was assessed using the Dr. Scratch abstraction and decomposition score (Table 5).
Figure 13. Scratch Challenge Starter Project.
Figure 14. Scratch Challenge Optimal Solution.
Figure 15. Sample Question From the Multiple-Choice Scratch Abstraction Test (the Correct Answer is D).
Projects were also manually analysed against 'completeness criteria' (an explanation of each is given below) as a measure of whether the project had been completed using an optimal solution. This is because Dr. Scratch can only measure whether a block has been used in a project, not whether it has been used properly.
Completeness criteria:
-
Correct custom block - A custom block with two inputs representing distance and degrees (may not be named correctly), containing a 'repeat' 'move' for the distance and a 'turn' for the direction.
-
Correct use of cloning - A single lamp post sprite that is cloned at the position of the cat sprite inside the custom block before the repeat (because the starter project has a lamp post sprite at the starting position.)
-
Complete movement - The cat sprite is animated around the map and reaches the shop as it does in the starter project.
-
Correct lamp post positions - All the lamp post sprites are in the same positions as they are in the starter project. They must appear in sequence (ideally as the cat sprite reaches them.)
Multiple-Choice Scratch Abstraction Test
The Scratch abstraction test was a 10-question multiple-choice assessment designed by the first author and used to supplement the Scratch challenge. The questions are on using custom blocks and cloning correctly in Scratch. Each question has four options with one correct answer. The test included questions on:
-
Identifying duplicated Scratch code that can be refactored using a custom block.
-
Identifying correct block names and inputs for duplicated code (Figure 15 shows a sample question).
-
Comparing Scratch scenes and figuring out which sprites should be cloned.
-
Identifying the block that can be used to get properties of a sprite.
-
Identifying the blocks used to clone sprites successfully.
Computational Thinking Test
The Computational Thinking test (CTt) (Román-González et al., 2017) was used as a measure of computational thinking. It aims to measure "the ability to formulate and solve problems by relying on the fundamental concepts of computing, and using logic-syntax of programming languages: basic sequences, loops, iteration, conditionals, functions and variables" (p. 4). The CTt contains 28 multiple choice questions that use visual arrows or blocks common in educational programming tools.
Ideally, the CTt would be combined with another computational thinking assessment that does not use programming syntax, such as Bebras (Dagiene & Stupuriene, 2016). However, this was not possible in this study due to school logistics and time constraints.
Artifact-Based Interviews
Artifact-based interviews (Brennan & Resnick, 2012) were used to establish whether participants had understood the rationale for using procedural abstraction in the Scratch challenge. The interviews were one-to-one with a researcher and took place the day after the assessment. They began with open questions about the participant's project, to see if they could explain why they had done something without prompting from the researcher, before progressing to more leading questions about custom blocks and cloning. They were asked about their project (what each of the blocks did and why they had used them), alternative approaches they considered, why they had/had not used custom blocks, why they had/had not used cloning, alternative scenarios in which they would use custom blocks or cloning, before finishing on similarities between the task and Pirate Plunder and general feedback on the game.
To select participants for the interviews, each intervention group was divided into three categories: correct solution, almost correct or interesting solution and no use of procedural abstraction. Five participants were selected from each category.
Game Analytics
Pirate Plunder produces analytics for player actions. Table 6 shows when and why this data is produced, with the overall aim of figuring out why the game is effective (RQ2) ("Rq2 - What Aspects of the Game Design Influence the Effectiveness of This Approach and Why?" section).
| Table 6. Pirate Plunder Analytics and Their Purpose. |
Procedure
All participants completed the Scratch baseline task and the CTt at pre-test. The Scratch baseline task took place in the school IT suite in class groups. Participants were introduced to the study and the assessment task. They were then given 40 minutes to produce a Scratch project to the assessment specification. The CTt was administered using tablets in a classroom after the group had completed the Scratch baseline task. Participants were given a maximum of 45 minutes to complete the test.
Class groups were then assigned to the intervention (Pirate Plunder) or active control conditions (spreadsheets and Scratch) for phase 1 of the study. Both phases were four weeks long with two sessions per week (30 minutes and 50 minutes), taking place in the school IT suite.
At mid-test, all participants did the Scratch challenge (as opposed to the Scratch baseline task), multiple-choice Scratch abstraction test and the CTt. The intervention group also completed a questionnaire and 15 of them were interviewed. Once again, the Scratch challenge took place in the IT suite with participants given 40 minutes to modify the starter project. Both the multiple-choice Scratch abstraction test and CTt (in that order) were then administered using tablets in the classroom. They were given a maximum of 15 minutes for the multiple-choice abstraction test and 45 minutes for the CTt.
The conditions were then crossed over so that the intervention group did spreadsheets and the two control groups from phase 1 did Pirate Plunder. At post-test, the intervention groups re-completed the Scratch challenge, multiple-choice Scratch abstraction test and the CTt, whilst the control group only did the multiple-choice test and the CTt (Figure 11 ). Thirty of the phase 2 intervention participants were then interviewed across the following two days.
Data Analysis
Hypotheses
The study had two hypotheses, both tested using the data from phase 1:
-
Pirate Plunder would perform better on the procedural abstraction measures in comparison with non-programming (spreadsheets) and programming (Scratch curriculum).
-
Pirate Plunder would improve scores on the CTt in comparison with the non-programming control group who were not doing explicit computational thinking activities.
The first hypothesis is tested using the Scratch challenge and multiple-choice Scratch abstraction test, using the Scratch baseline task as a covariate on the Scratch challenge between-groups comparison. The second hypothesis is tested using a between-groups comparison between the CTt scores from pre-to mid-test.
Statistical Methods
To test the first hypothesis, we use a one-way ANCOVA to compare the mid-test Scratch challenge Dr. Scratch abstraction and decomposition scores for each group using the Scratch baseline task Dr. Scratch abstraction and decomposition scores as a covariate, to control for variance in baseline ability. In addition to a one-way ANOVA to compare the multiple-choice Scratch abstraction test mid-test scores.
For the second hypothesis, we use a one-way ANOVA to measure for a between-groups difference, then a series of independent samples t-tests to test for significant pairwise comparisons.
Research Questions
RQ1 is addressed using the procedural abstraction measures (Scratch challenge and multiple-choice Scratch abstraction test) and is directly linked to the first hypothesis. RQ2 is addressed using the Pirate Plunder game analytics.
Phase 2
In phase 2 we would expect to see an improvement on the procedural abstraction measures for the groups who played Pirate Plunder. Due to the crossover design, there was not a genuine control group for phase 2 (as the control group had already been exposed to the intervention at this stage). As such, this section uses within-group comparative measures from before and after Pirate Plunder, instead of comparisons between groups. The groups are identified using their phase 1/phase 2 learning content (e.g. Pirate Plunder/spreadsheets).
Scratch Challenge
Both phase 2 intervention groups improved significantly from mid-to post-test: spreadsheets/Pirate Plunder (t 28 = 5.52, p < .001, d = 1.44) and Scratch/Pirate Plunder (t 25 = 8.76, p < .001, d = 1.44). The phase 2 control group did not re-complete the assessment at post-test due to school limitations.
Multiple-Choice Scratch Abstraction Test
A one-way ANOVA showed that there was no significant difference between groups: F 2,71 = 2.21, p = .12, η2 = .059. In addition, using a paired samples t-test, only the Scratch/Pirate Plunder group changed (in this case, improved) significantly from mid-to post-test: t 26 = 2.14, p = .042, d = 0.47.
Computational Thinking Test
A one-way ANOVA showed a significant difference between groups, F 2, 84 = 4.49, p = .014, η2 = .097. A paired samples t-test showed that the Pirate Plunder/spreadsheets group declined significantly: t 27 = 2.87, p = .008, d = 0.38.
Scratch Challenge Completeness Criteria After Playing Pirate Plunder
The Scratch challenge completeness criteria ("Scratch Challenge" section) were used to support the Dr. Scratch abstraction and decomposition scores. Table 9 shows the number of participants that met each criterion in their Scratch challenge projects before and after their Pirate Plunder intervention. There were significant improvements for the phase 2 intervention groups from mid-to post-test in using the correct custom block (spreadsheets/Pirate Plunder, t 28 = 6.84, p < .001, d = 1.80 and Scratch/Pirate Plunder, t 25 = 9.21, p < .001, d = 2.56) and the correct use of cloning (spreadsheets/Pirate Plunder, t 28 = 2.42, p = .023, d = 0.63 and Scratch/Pirate Plunder, t 25 = 2.13, p = .043, d = 0.58).
| Table 9. Scratch Challenge Completeness Criteria Before and After Pirate Plunder Intervention. |
Pirate Plunder Player Performance
Table 10 shows the Pirate Plunder player performance for each group. This was judged using the number of challenges completed (maximum of 40) and overall stars collected (maximum of 120). The average stars collected on each level (maximum of 3) is also given. One-way ANOVAs showed no significant difference between the three groups for challenges completed (F 2,87 = 0.81, p = .447, η2 = .018) or stars collected (F 2,87 = 1.13, p = .329, η2 = .025).
| Table 10. Descriptive Statistics for Pirate Plunder Player Performance. |
Artifact-Based Interview Observations
When asked if they could give another example in which they would use a custom block in Scratch, most participants (29/45, 64.4%) gave examples situated in the context that they had learnt to use procedural abstraction (i.e. involving moving and turning a sprite) (category A). However, 12 participants (26.7%) were able to apply procedural abstraction to theoretical scenarios outside of Pirate Plunder or could explain general rules when using procedural abstraction (category B). For example, one participant said that they could use custom blocks and cloning when creating a bowling game. Even going as far as to question whether cloning would be appropriate or not due to the way it works:
Researcher: "Can you give me another example of where you'd use a custom block in Scratch?"
Participant: "You could create a bowling game and you could input the amount of power the ball would move, so you could determine how far it would go, or you could use it for some sort of game where you'd throw or catapult something. So, you could change at different moments how far it would go."
The remaining four participants (8.9%) struggled to apply procedural abstraction to any scenario (even if they had used it in the Scratch challenge and/or the game) (category C). These categories (when scored 1-3) correlate significantly with participant CTt pre-test scores, r 45 = .34, p = .021. Table 11 shows relevant descriptive statistics.
| Table 11. Average CTt Pre-Test Score for Interview Observation Category. |
Discussion
Rq1 – Can a Game-Based Learning Approach Be Used to Teach Primary School Children to Use Procedural Abstraction in Scratch projects?
The phase 1 results support the first hypothesis: children's improvements on measures of procedural abstraction after playing Pirate Plunder were greater in comparison with the control groups. In phase 2, the children that played Pirate Plunder significantly improved their scores on the Scratch challenge.
In addition, the Scratch challenge completeness criteria show that the game was effective in getting children to use procedural abstraction correctly. But, less so than the Dr. Scratch abstraction and decomposition scores indicate. Sixty-one participants achieved a Dr. Scratch abstraction and decomposition score of 2 or 3 (for using custom blocks or cloning) on the Scratch challenge after playing Pirate Plunder. Yet, only 34 of these produced the correct custom block and nine produced the correct cloning solution. This is a weakness of using Dr. Scratch as a measure in a study such as this because it does not account for whether a block or functionality has been used correctly.
This addresses RQ1 by showing that primary school children (age 10 and 11) can be taught to use procedural abstraction in Scratch projects using a game-based learning approach.
Computational Thinking
The results of the Computational Thinking test support hypothesis 2, that Pirate Plunder improved participants' computational thinking compared to the non-programming control after phase 1 of the study. Yet, these results were not repeated after the crossover. The decline in the Pirate Plunder/spreadsheets group in phase 2 is likely because the participants were doing the same assessment for the third time and had lost some motivation to complete it properly.
There are, however, issues with using programming-based assessments as measures of computational thinking, as they do not consider the wider use of computational thinking in problem-solving (Kazimoglu et al., 2011).
Rq2 – What Aspects of the Game Design Influence the Effectiveness of This Approach and Why?
In this section, we give four reasons why Pirate Plunder was effective that are supported by the game analytics ("Game Analytics" section).
Motivating Players Using Restrictive Success Conditions
For players to experience flow (Csikszentmihalyi, 1990), games must maintain the balance between the challenge of the levels and ability of the player, whilst also providing clear goals and immediate and accurate feedback. Pirate Plunder was effective in keeping players motivated using these flow enablers. On average, players completed 82.8% (33.12/40) of the challenge levels by the end of the intervention. Additionally, they spent 80.07% of their time playing through the levels (as opposed to time spent on the shop, class screen or level select).
As described in "Concept Scaffolding" section, restrictive success conditions such as block limits (the number of blocks a player can use to complete a level) and required block validation (completing the level using the block linked to that challenge) were used to force the player into producing optimal solutions. This was an effective approach, despite observations that players often tried to circumvent program restrictions. Players had an average score of 2.93/3 stars per challenge level, meaning that they were using the correct blocks and producing optimal solutions (in terms of block count) 97.66% of the time. In addition, players tended not to reattempt levels once they had completed them (the average number of total challenge reattempts per player was 3.1), meaning that the high average scores were achieved the first time the player completed the level.
Effective Customisation System
Customisation is a powerful motivator in both learning (Cordova & Lepper, 1996) and games (Turkay & Adinolf, 2015). In Pirate Plunder, it is used together with a reward system, where players collect coins for completing levels successfully ("Customisation" section). The game analytics showed that this was an effective design strategy. Players spent 79.36% of their total coin earnings on purchasing items and customising their avatars. Players also purchased items regularly throughout gameplay. Across the eight sessions that each player was given, players purchased an average of 3.14 items and spent 84.33 coins per session (Figure 19 ). This shows that the coin rewards for playing the game were an effective method of motivating players to continue through the learning content, as was the unlocking of items as the player progressed through the challenge levels. This was the case even though avatar items do not give the player an advantage in the game and are for aesthetics only.
Figure 19. Bar Chart Showing the Average Number of Coins Spent in Each Session.
Allowance for Different Levels of Prior Knowledge
Papert (1980) stated that educational programming tools should have a low floor (easy access) and a high ceiling (vast potential) to be inclusive for all learners. The Pirate Plunder learning trajectory ("Concept Scaffolding" section) follows these principles, introducing basic Scratch blocks and functionality first. This was an effective approach in lowering the barrier of entry for the game. Players who achieved less than the mean score (14.85) on the CTt at pre-test still completed 74.9% (29.97/40) of the challenges (7.9% lower than the mean). Additionally, the game also had a suitably high ceiling. None of the 20 participants who completed all 40 levels introducing the learning content were able to complete all eight of the 'general levels' ("Concept Scaffolding" section) that become available after level 40 is complete. The CTt pre-test is used here because it is a more general measure of programming ability than the Scratch-based assessments.
Effective Difficulty Curve
Difficulty scaling is a fundamental part of game design (Aponte et al., 2009). Despite not being adjusted dynamically on a per-player basis, the Pirate Plunder difficulty progression fits with tension-resolution cycles linked to player enjoyment: when a new concept is introduced, the player initially feels tension until they gain an understanding of the concept (resolution). Figure 20 shows that the difficulty curve was effective in introducing new concepts without them being too difficult, with players completing introductory levels quickly using few attempts. There is a jump for the last 'loops' challenge (level 15) (Figure 5 ), then a larger jump when procedures are introduced (level 19). The dip after this introduction shows that the 'custom block' and 'inputs' levels introduced further complexity successfully without being too difficult.
Figure 20. Line Graph of the Average Attempts per Challenge and the Average Time in Minutes per Challenge (Both per Level Completion).
Recommendations
In this section, we draw out three recommendations for educational game design from our research.
Using Learning Trajectories and Restrictive Success Conditions to Introduce Complex Learning Content
Educational games allow for granular management of when and how learning content is introduced. Structuring this effectively is an important part of educational game design. The results of this study and the reasons the game was effective ("Rq2 - What Aspects of the Game Design Influence the Effectiveness of This Approach and Why?" section) show that learning trajectories, such as the one we have used to introduce procedural abstraction, are an effective method of introducing conceptually difficult learning content. By designing levels around this, the learner can use the concept without instruction from the teacher, because the learning content is introduced and reinforced by the game. Restrictive success conditions on levels, such as block limits and validation conditions, are important in getting players to use taught functionality correctly. This is particularly important in primary education, where teachers often lack theoretical and technical knowledge of computing.
Increasing Learner Investment Through Customisable Avatars
Motivating learners to progress through the content is vital in educational games. Particularly when the novel learning content is introduced in the latter parts. The regularity and amount of player spending ("Effective Customisation System" section), combined with the high percentage of maximum scores on Pirate Plunder levels ("Motivating Players Using Restrictive Success Conditions" section), shows that players were invested in their avatars and this served as a good motivator for success. We, therefore, recommend that avatars, where the player can upgrade them using currency or rewards earnt playing the game, are a good method of increasing learner investment. This combines the extrinsic motivator of purchasing items for the avatar with the intrinsic motivator of collecting the maximum number of coins possible by completing levels.
In earlier versions of Pirate Plunder, we placed a small cost on saving changes to an avatar to ensure that players did not spend too much time customising their avatar early in the game (when they have a limited number of coins). However, these fears were unfounded, and we removed this for the study reported in this article.
Improving Evaluations of Educational Games
One of the strengths of this work is that Pirate Plunder was evaluated against existing curricula that are used in English primary schools, including both programming (Scratch) and non-programming (spreadsheets) lesson plans, using a range of quantitative and qualitative assessments.
Two large literature reviews of evaluations of educational games in the classroom (Hainey et al., 2016; Petri & Gresse von Wangenheim, 2017) state that whilst there are many examples of strong study designs, either randomised controlled trials or quasi-experimental studies, there are a larger number of studies that use simple, ad-hoc research designs, subjective feedback via questionnaires and small sample sizes to evaluate educational games.
The combination of the range of educational programming tools available ("Educational Programming Tools" section) and the increasing pressure to deliver computer science learning content, mean that robust evaluations of programming games are essential. Otherwise, it is difficult to know whether these tools are effective in a real-world classroom environment.
Limitations
Potential Experimenter Bias
As the lead author delivered both the intervention and control group content, there is a possibility that experimenter bias could have played a role in the outcomes of the study. Ideally, the control group content would have been delivered by a teacher blind to the hypothesis. We aim to address this issue in future empirical work ("Future Work" section).
Situated Learning
The participants used procedural abstraction in a situated context, namely 'moving' and 'turning' in Scratch, with the majority then only able to explain procedural abstraction within that context ("Artifact-Based Interview Observations" section). However, the finding that they can learn to use procedural abstraction in any context is important and one that to our knowledge has not been demonstrated by primary school children before (RQ1). In addition, there were instances of higher scoring participants (on the CTt at pre-test) being able to explain how they could apply procedural abstraction in other Scratch projects after playing Pirate Plunder. We aim to address this issue in future work by having participants use Scratch to produce projects using procedural abstraction after playing Pirate Plunder.
Conclusions
In conclusion, we have addressed RQ1 in showing that Pirate Plunder can be used to teach primary school children (age 10 and 11) to use procedural abstraction in Scratch projects. This is a significant finding and indicates that procedural abstraction can be part of computer science curricula for this age group, supporting the results of Kalas and Benton (2017).
Using the game analytics, we have then addressed RQ2 by giving reasons why the game was effective. Firstly, because players were motivated to progress through the content using our learning trajectory and restrictive success conditions. Secondly, customisation was effective in motivating players to continue playing the game. Thirdly, the game allows for players with less prior knowledge to still progress through the game. Finally, the difficulty curve is effective in introducing new concepts. This then led into three recommendations for designing programming games to support computer science knowledge: using learning trajectories and restrictive success conditions to introduce complex learning content, increasing learner investment through customisable avatars and improving the evaluation of educational games.
The success of Pirate Plunder shows that game-based learning can play a key role in supporting and delivering computer science content in primary education. We hope that our recommendations will be used by game designers to improve the delivery of learning content in educational games.
Future Work
Future empirical work will address the limitations of this study. Firstly, to conduct studies in multiple schools to confirm our findings with a more generalisable sample. Secondly, to have the control group content delivered by a teacher or researcher blind to the hypothesis, to remove any potential experimenter bias. We also aim to evaluate Pirate Plunder against a traditional computer science curriculum that introduces procedural abstraction, as opposed to standard computing curricula. This will allow us to evaluate whether the game is effective in comparison with another form of instruction. Further to this, we would like to explore whether the procedural abstraction skills learnt in Scratch (using Pirate Plunder) transfer to other Scratch projects, as indicated by the artifact-based interviews and then onto text-based languages, whilst investigating how this transfer can be effectively mediated.
In terms of development, we aim to extend Pirate Plunder to include other computer science concepts that novices struggle with, such as variables and conditionals (Grover & Basu, 2017). The game would also be updated for use alongside Scratch 3, which due to being a native web application, now allows for easier extension and project analysis (e.g. Stahlbauer et al., 2019).
Declaration of Conflicting Interests Funding
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
The authors received no financial support for the research, authorship, and/or publication of this article.
Google Scholar
Google Scholar
Google Scholar
Google Scholar | Crossref | Medline
Google Scholar | Crossref
Google Scholar
Google Scholar
Google Scholar | Crossref
Google Scholar
Google Scholar | Crossref
Google Scholar | Crossref | ISI
Google Scholar
Google Scholar | Crossref
Google Scholar
Google Scholar | Crossref
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar | Crossref
Google Scholar | Crossref
Google Scholar | Crossref
Google Scholar
Google Scholar
Google Scholar | Crossref
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar | Crossref
Google Scholar | Crossref
Google Scholar | SAGE Journals | ISI
Google Scholar
Google Scholar
Google Scholar | Crossref
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar | Crossref
Google Scholar | Crossref
Google Scholar | Crossref
Google Scholar
Google Scholar | Crossref
Google Scholar | Crossref
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar | Crossref
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar | Crossref
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Google Scholar
Cs First Game Design Sprite Are Disappearing
Source: https://journals.sagepub.com/doi/full/10.1177/0735633120932871
Posted by: shawuponce.blogspot.com
0 Response to "Cs First Game Design Sprite Are Disappearing"
Post a Comment