EdTech Archives EdTech Archives International Consortium for Innovation and Collaboration in Learning Engineering (ICICLE) 2024 Conference Proceedings: Solving for Complexity at Scale

Using a Nested Learning Engineering Methodology to Develop a Team Dynamic Measurement Framework for a Virtual Training Environment

Scotty D. Craig, Kamala Avancha, Parkhi Malhotra, Jamie C. Gorman, Vipin Verma, Robert Likamwa, Kevin Gary, Randall Spain, & Benjamin Goldberg

Abstract

Evaluating team dynamics and shifts in these dynamics is essential when designing adaptive training in synthetic training environments. In this proceedings presentation, we discuss how we utilized an overarching Learning engineering process to develop a team dynamics framework within the Tactical Combat Casualty Care (TC3) synthetic training and also to monitor and improve team dynamics throughout the training. The goal of the framework is to assess and improve team members’ ability to recover and reorganize from unexpected but significant disturbances or system failures.

Introduction

Arizona State University (ASU) collaborated with the US Army Soldier Center Simulation and Training Technology Center (STTC) to address a pervasive challenge of monitoring and training team dynamics. A key goal of addressing this challenge from the U. S, Army’s view is to create training as an example of their Synthetic Training Environment (STE) initiative within a broadly used application domain. While live simulation training can be effective, it is often expensive, time-consuming, and dangerous (Levy et al. 2013). Simulations and virtual systems have a long history of providing effective training (Swezey & Andrews, 2001) with decreased danger of injury during training (Andrews & Craig, 2015). STE training has been shown to be effective for team training (Siegle et al., 2021); however, it is recommended that this type of environment be used in areas with set content areas that do not require redevelopment that could cause rapidly increasing expense (Siegle et al., 2020). The project’s goal was to develop an STE training experience(i.e., immersive virtual reality) for Tactical Combat Casualty Care training (Craig et al., 2024). The system monitors team dynamics and helps improve team members’ ability to recover and reorganize from unexpected, yet significant, disturbances or system failures involved in combat lifesaving missions (Avancha et al., 2024). The Tactical Combat Casualty Care (TC3, Butler, n.d.) area was selected based on conversations with U. S. Army stakeholders and follow-up interviews with subject matter experts in the TC3 area. Effective teamwork, as noted by Marlow and colleagues, is a key factor in improving patient outcomes and minimizing medical errors, particularly in high-stress, mass-casualty situations (Marlow et al.,2018). Furthermore, a comprehensive curriculum, which incorporates decision-making under high-pressure conditions, is essential for providing tactical medical support in fast-paced combat scenarios (Milham et al., 2017).

The current paper describes the development of a team dynamics measurement framework, one smaller aspect of our larger system, to provide an example of the application of nested learning engineering process. Our team consisted of learning scientists, human factors specialists, a Project Manager, Software Engineers, Virtual environment developers, and military researchers with subject matter expertise on TC3, combat lifesaving, and military training.

Learning Engineering Process

The development of our system adopted a broader Learning Engineering Process (Goodell & Kolodner, 2023; Kessler et al, 2023; Craig et al., 2023) which consists of four phases (see Figure 1): Challenge (understanding the problem in context), Creation (constructing an inclusive solution that meets the needs of end users and addresses the problem), Implementation (observing the solution in progress with data collected on impact), and Investigation (analyzing data to understand the impact of the solution on the challenge and the various stakeholders). However, any large-scale project will require smaller cycles of the process. We refer to these inner cycles as nested cycles for ease of description (Totino & Kessler, 2024). These mainly occur during the challenge phase, with multiple cycles needed to understand the challenge before moving forward, and in the creation phase, where multiple pieces of the solution need to be created. However, nested cycles could also occur during the investigation phase for refinement of analysis to better inform the next iteration. It would be less common for nested cycles to occur during implementation because the enactment of the solution needs to be consistent so that the data collected can be applied to the interaction of the implemented solution to the challenge (Czerwinski et al., 2022).

As highlighted in the opening of this paper, our initial challenge phase of the project consisted of multiple nested cycles. We had a series of conversations (informal unstructured interviews) with our military partners' representatives to understand their training needs, their needs around how the training would be built, and the targeted end users. Further conversations were held with subject matter experts to improve understanding of the content domain and the end users for the training. From these conversations, our educational scientists reviewed the initial needs and provided best practices for virtual training.

Figure 1

Learning Engineering process and nested iteration for creation phase

A Nested Cycle within the Creation Phase

The remainder of this paper provides a detailed case study of a nested cycle within the creation phase. This nested cycle is an example of the nested Learning Engineering Process Model which has micro loops of the learning engineering process within one of the four phases of the larger process (Kessler & Totino, 2023). These phases allow for development and testing of individual part of the solution (Totino & Kessler, 2024).This cycle will focus on the identification of the learning content on TC3. The primary method used in this cycle was a Hybrid Cognitive Task Analysis (hCTA, Nehme et al., 2006). The hCTA is a specialty form of Cognitive Task Analysis (Clark et al., 2008) used to generate information and functional requirements using a scenario description and an enumeration of high-level mission goals. It has four steps: 1) generating a scenario task overview, 2) generating an event flow diagram, 3) creating decision ladders for decision points, and 4) generating a situational awareness requirement (Nehme et al., 2006; Macbeth et al., 2012). This process requires reviews of written materials, interviews with subject matter experts, creation of training material scenarios, and validation of final materials with SME.

Designing Training Content

Our initial challenge during the creation phase was to design a scenario that reflects Tactical Combat Casualty Care team training within a STE (Goldberg et al., 2021). The goal was to emphasize team dynamics, collaboration, and performance under stress all while replicating the technical aspects of Tactical Combat Casualty care (Avancha et al., 2024). We began with a deep dive of the available team training literature for TC3 (Townsend et al., 2016; Madrid et al., 2020; Milham et al., 2017; Ross et al., 2016; Johnston, 2019; Gerhardt, 2012). Literature review highlighted the role of immersive, simulation-based training in strengthening TC3 training through collaborative team interactions and enhanced overall performance.

Creation

Following our literature review, we engaged with the existing computer based TC3 training to gain an understanding of current training resources available. applied Hybrid Cognitive Task Analysis methodology, a structured approach that helped us break down the complex tasks into actionable steps required for team based performance. This process involved close examination of each role within a TC3 team, identifying the specific skills, competencies, and communication patterns necessary for the team to function effectively. To ensure the scenario we developed captured the realities of TC3 operations, we conducted a series of semi-structured interviews with subject matter experts (SMEs), including experienced military trainers, squad leaders, STTC training designers, and learning engineers. Interviews were conducted virtually through Microsoft Teams where the SMEs guided us through the different TC3 roles, operations, areas within the CCP, and the processes involved in casualty triage and support within the CCP. These interviews provided valuable insights into Casualty Collection Point (CCP) operations, team structures, and the current training available to Combat life savers. This step was crucial to ensure that the training being built reflects realistic interactions and task requirements. Through this, we identified gaps in existing training, which informed the design of our scenario and the necessary team competencies.

hCTA Part 2

Using this information from SME interviews and hCTA, we developed an event flow diagram that detailed task sequences and situational awareness across team members. Additionally, we introduced critical perturbations, which are unexpected but significant disruptions—such as disorganization and unplanned changes in team leadership or environmental challenges, to assess and improve team adaptability.

Implementation & Investigation

Following the initial design, we conducted multiple rounds of feedback from SMEs. In the first review, SMEs focused on validating the accuracy of the CCP design, team roles, and operational flow within the scenario. Based on their input, we refined the training scenario and reintroduced it for a second round of review. In this phase, we focused on the fine-tuning of the non-player character (NPC) scripts, ensuring that the interactions and dialogue reflected real-life battlefield conditions. The final round of review provided detailed feedback on the remaining discrepancies and allowed us to make minor adjustments to enhance the trainee experience.

Based on the SME feedback and our iterative designs during investigation, we finalized a comprehensive pilot scenario for a CCP in an urban setting, which included various perturbations to challenge trainees. The scenario was then passed to the Software Engineering and Virtual Environment teams for technical development.

Finally, the scenario involved three Combat Life Saver (CLS) trainees operating in a high-stress CCP setting, supported by NPCs representing various roles. The scenario, divided into two 15-minute sessions with debriefing breaks, aimed to test the trainees’ ability to manage casualties, communicate effectively, and adapt under pressure, ultimately refining their ability to operate as a cohesive, adaptive team (see Fig.2).

Figure 2

Final Scenario that was developed

Conclusion

In conclusion, the current nested cycle provided a foundation to develop this solution using a multiuser VR system utilizing wireless tethered headsets connected to the client computers with a central server that tracks and manages game states. After the development, we plan on an implementation phase nested cycle and conduct an empirical heuristic evaluation. Our heuristic evaluation will examine usability and effectiveness through established evaluation principles. Finally, we will enter the Investigation phase to validate metrics and determine the impact of the training system developed. Additionally, we will assess whether these Learning Engineering cycles should be repeated to enhance our VR training in future.

By following the Learning Engineering process of “Challenge, Create, Implement, and Investigate”, we were able to develop a highly realistic and effective TC3 training within the Synthetic Training Environment. We began by addressing the challenge of accurately simulating team dynamics and performance under the stress of Tactical Combat Casualty Care. This helped us create a robust initial pilot scenario informed by Hybrid Cognitive Task analysis, Subject Matter Expert interviews, and iterative design feedback. Once the design was validated through multiple rounds of SME reviews, we proceeded to develop it on the Virtual Reality platform. Aspects such as perturbations were integrated to enhance immersive training experience. Finally, our goal is to investigate the training effectiveness in terms of team dynamics and adaptability and optimize the learning environment. This process allowed us to design a training experience that enhances trainees’ ability to adapt and perform as a cohesive team in high pressure situations.

Acknowledgments

The research described herein has been sponsored by the U.S. Army Combat Capabilities Development Command under cooperative agreement W912CG-23-2-0002. The statements and opinions expressed in this article do not necessarily reflect the position or the policy of the United States Government, and no official endorsement should be inferred.

References

  1. Andrews, D. H., & Craig, S. D. (Eds.). (2015). Readings in training and simulation (Vol. 2): Research articles from 2000 to 2014. Santa Monica, CA: Human Factors and Ergonomics Society.
  2. Avancha, K., Malhotra, P., Gorman, J. C., Verma, V., Spain, R., Goldberg, B., & Craig, S. D. (2024, December). Development of Team Dynamic Measurement Framework for Adaptive Teams. In the Proceedings of Interservice/Industry Training, Simulation and Education Conference 2024 (Paper # 24324). National Training & Simulation Association.
  3. Butler, F., MD & Center For Army Lessons Learned. (n.d.). Tactical Combat Casualty Care Handbook (By M. F. Pappal COL, R. St. Jean CPT, W. Engle MAJ, & J. C. Fruendt M. D.). https://usacac.army.mil/sites/default/files/publications/17493.pdf
  4. Clark, R. E., Feldon, D. F., Van Merrienboer, J. J., Yates, K. A., & Early, S. (2008). Cognitive task analysis. In Handbook of research on educational communications and technology (pp. 577-593). Routledge.
  5. Craig, S. D., Gary, K., Gorman, J. C., Verma, V., & LiKamWa, R. (2024, August). A Synthetic Training Environment for Assessing Changes in Team Dynamics with the Generalized Intelligent Framework for Tutoring. In A. M. Sinatra (Ed.), Proceedings of the 12th Generalized Intelligent Framework for Tutoring (GIFT) Users Symposium (GIFTSym12) (p. 97-103). U. S. Army Combat Capabilities Development Command-Soldier Center.
  6. Craig, S. D., Goodell, J., Czerwinski, E., Lis, J., & Roscoe, R. D. (2023). Learning Engineering Perspectives for Supporting Educational Systems. Proceedings of the Human Factors and Ergonomics Society Annual Meeting, 67(1), 304-309. https://doi.org/10.1177/21695067231192886
  7. Czerwinski, E., Goodell, J., Ritter, S., Sottilare, R., Thai, K. P., & Jacobs, D. (2022). Learning engineering uses data (part 1): Instrumentation. In J. Goodell & J Kolodner (Eds.). Learning Engineering Toolkit (pp. 153-174). Routledge.
  8. Gerhardt, R., Mabry, R., DeLorenzo, R., & Butler, F. (2012). Fundamentals of combat casualty care. Chapter 3, Combat Casualty Care: Lessons Learned from OEF and OIF, Borden Institute. US Army Medical Department Center and School. USA: Pelagique LLC.
  9. Goldberg, B., Owens, K., Hellman, K., Robson, R., Blake-Plock, S., Hoffman, M., & Gupton, K. (2021). Forging Competency and Proficiency through the Synthetic Training Environment with an Experiential Learning for Readiness Strategy. In Proceedings of the Interservice Industry, Training, Simulation, and Education 2021 Conference. NTSA. https://www.xcdsystem.com/iitsec/proceedings/index.cfm?Year=2021&AbID=97151&CID=862
  10. Goodell, J., & Kolodner, J. (2023). Learning engineering toolkit: Evidence-based practices from the learning sciences, instructional design, and beyond. Routledge.
  11. Johnston, J.H. (2019). Team Performance and Assessment in GIFT - Research recommendations based on Lessons Learned from the Squad Overmatch Research Program. Army Research Laboratory, Human Research and Engineering Directorate.
  12. Kessler, A., Craig, S. D., Goodell, J., Kurzweil, D., & Greenwald, S. W. (2023). Learning engineering is a process. In J. Goodell & J Kolodner (Eds.). Learning Engineering Toolkit (pp. 29-46). Routledge.
  13. Kessler, A. & Totino, L. (2023, July 24-26). Reality is Much More Complex: Understanding and Exploring Nested Learning Engineering Processes. IEEE ICICLE Learning Engineering Conference 2023, Pittsburgh, PA, United States.
  14. Levy, M., Koch, R. W., & Royne, M. B. (2013). Self-reported training needs of emergency responders in disasters requiring military interface. Journal of Emergency Management,11(2), 143–150.
  15. Macbeth, J. C., Cummings, M. L., Bertuccelli, L. F., & Surana, A. (2012, September). Interface design for unmanned vehicle supervision through hybrid cognitive task analysis. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting (Vol. 56, No. 1, pp. 2344-2348). Sage CA: Los Angeles, CA: SAGE Publications.
  16. Madrid, R., Pike, W., Cummings, P., & Barnieu, J. (2020). Addressing Tactical Combat Casualty Care in Synthetic Training Environments. Proceedings of the 2020 Interservice/Industry Training, Simulation, and Education Conference (I/ITSEC). Arlington, VA: National Training and Simulation Association.
  17. Marlow, S. L., Bedwell, W. L., Zajac, S., Reyes, D. L., LaMar, M., Khan, S., Lopreiato, J. L., & Salas, E. (2018). Multiple Patient Casualty Scenarios: A Measurement Tool for Teamwork. Simulation in Healthcare, 13(6), 394–403.
  18. Milham, L. M., Phillips, H. L., Ross, W. A., Townsend, L. N., Riddle, D. L., Smith, K. M., Butler, P. V., Wolf, R. J., Irizarry, D. J., Hackett, M. G., & Johnston, J. H. (2017). Squad-level training for Tactical Combat Casualty Care: Instructional approach and technology assessment. The Journal of Defense Modeling and Simulation: Applications, Methodology, Technology, 14(4), 345–360. https://doi.org/10.1177/1548512916649075
  19. Nehme, C. E., Scott, S. D., Cummings, M. L., & Furusho, C. Y. (2006, October). Generating requirements for futuristic hetrogenous unmanned systems. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting (Vol. 50, No. 3, pp. 235-239). Sage CA: Los Angeles, CA: Sage Publications.
  20. Ross, W. A., Johnston, J. H., Riddle, D., Phillips, C. H., Townsend, L., & Milham, L. (2016). Making sense of cognitive performance in small unit training. In Foundations of Augmented Cognition: Neuroergonomics and Operational Neuroscience: 10th International Conference, AC 2016, Held as Part of HCI International 2016, Toronto, ON, Canada, July 17-22, 2016, Proceedings, Part II 10 (pp. 67-75). Springer International Publishing.
  21. Siegle, R. F., Roscoe, R. D., Schroeder, N. L., & Craig, S. D. (2020). Immersive Learning Environments at Scale: Constraints and Opportunities. Proceedings of the Human Factors and Ergonomics Society Annual Meeting, 64(1), 1165-1169. https://doi.org/10.1177/1071181320641278
  22. Siegle, R. F., Cooke, N. J., Schroeder, N. L., Li, S., & Craig, S. D. (2021). Scaling team training: Using virtual worlds to support learning in massive open online courses. Proceedings of the Human Factors and Ergonomics Society Annual Meeting, 65(1), 566-570. https://doi.org/10.1177/1071181321651204
  23. Swezey, R. W., & Andrews, D. H. (2001). Readings in training and simulation: A 30-year perspective. Human Factors and Ergonomics Society.
  24. Townsend, L., Milham, L., Riddle, D., Phillips, C. H., Johnston, J., & Ross, W. (2016). Training Tactical Combat Casualty Care with an Integrated Training Approach. In D. D. Schmorrow & C. M. Fidopiastis (Eds.), Foundations of Augmented Cognition: Neuroergonomics and Operational Neuroscience, 9744(253–262). Springer International Publishing. https://doi.org/10.1007/978-3-319-39952-2_25
  25. Totino, L., & Kessler, A. (2024). “Why did we do that?” A Systematic Approach to Tracking Decisions in the Design and Iteration of Learning Experiences. The Journal of Applied Instructional Design, 13(2) https://doi.org/10.59668/1269.15630