MEEGA+
MEEGA+[1][2] is a method to evaluate the quality of educational games used to teach computing/software engineering in terms of player experience and usability from the viewpoint of the learners ans instructors.
It is composed of an evaluation model (MEEGA+ Model) defining quality factors to be evaluated through a standardized measurement instrument, and a scale, which classifies the evaluated game according to its quality level, and a process (MEEGA+ Process) defining phases, activities and work products, guiding researchers in how to plan, execute and analyse the results of game evaluations.
A complete description of the MEEGA+ Method is available here.
The MEEGA+ Model[edit]
Improving the initial version of the MEEGA model[3], the MEEGA+ model has been developed by systematically decomposing the evaluation goal into measures and defining a measurement instrument.
It provides a ready-to-use kit (including a questionnaire and analysis spreadsheet) to game creators, instructors and researchers to evaluate the quality of educational games in and effective and efficient way in practice.
As it is applied in a one-shot post-test research design it can be used quickly in an non-intrusive way after the application of any kind of game (digital or non-digital). All materials are available here.
Decomposition of the factors evaluated in the MEEGA+ Model:
- Player experience:
- Focused Attention
- Fun
- Challenge
- Social Interaction
- Confidence
- Relevance
- Satisfaction
- Perceived Learning
- Usability:
- Aesthetics
- Learnability
- Operability
- Accessibility
The MEEGA+ Process[edit]
The MEEGA+ process provides a systematic support, detailing steps and interrelated activities, guiding researchers and instructors in the conduction of educational game evaluations.
Phases and activities of the MEEGA+ process:
- Scoping
- Select the object of study (educational game)
- Define the evaluation goal
- Planiing
- Specify the evaluation context
- Define the evaluation schedule
- Prepare the data collection instrument
- Obtain approval from the Human Research Ethics Committee (optional)
- Execution
- Prepare the execution of the evaluation
- Obtain participants' consent (optional)
- Execution of the game
- Collect the data
- Check the validity of the data collected
- Analysis
- Prepare collected data for analysis
- Check the descriptive statistics
- Classify the quality level of the game
- Interpret the data
- Discuss the results
- Presentation
- Report the evaluation
References[edit]
- ↑ Petri, Giani; Gresse von Wangenheim, Christiane; Borgatto, Adriano Ferreti (2017-12-23), "MEEGA+, Systematic Model to Evaluate Educational Games", Encyclopedia of Computer Graphics and Games, Springer International Publishing, pp. 1–7, doi:10.1007/978-3-319-08234-9_214-1, ISBN 9783319082349
- ↑ Petri, G.; Gresse von Wangenheim, C.; Borgatto, A. F. MEEGA+: A Method for the Evaluation of Educational Games for Computing Education. INCoD/GQS.05.2018.E (July/2018)
- ↑ Savi, Rafael; Wangenheim, Christiane Gresse von; Borgatto, Adriano Ferreti (September 2011). A Model for the Evaluation of Educational Games for Teaching Software Engineering. 2011 25th Brazilian Symposium on Software Engineering. IEEE. doi:10.1109/sbes.2011.27. ISBN 9781457721878. Search this book on
This article "MEEGA+" is from Wikipedia. The list of its authors can be seen in its historical and/or the page Edithistory:MEEGA+. Articles copied from Draft Namespace on Wikipedia could be seen on the Draft Namespace of Wikipedia and not main one.