Testing the Efficacy of an Artificial Intelligence Real-Time Coaching SystemSystemSimulatioTraining of Medical Students

Sponsor
McGill University (Other)
Overall Status
Completed
CT.gov ID
NCT05168150
Collaborator
(none)
98
1
3
3.9
25.3

Study Details

Study Description

Brief Summary

Background:

Trainees learn surgical technical skills through apprenticeship model working closely with surgeons and given increased responsibility in patient cases under expert supervision. However, factors such as surgeons' busy schedule, number of available patient cases, patient safety and lack of objectivity and standardization in training pose strong limitations. Virtual reality surgical simulators integrated with artificial intelligence (AI) systems provide a standardized realistic simulation environment and detailed performance data that allows accurate quantitation of surgical skills and tailored feedback. These platforms make repetitive practice of surgical skills possible in a risk-free environment.

The Intelligent Continuous Monitoring System (ICEMS), a deep learning application integrated in NeuroVR simulation platform, was developed to assess surgical performance continuously in 0.2 second intervals and provide coaching and risk detection. Although a predictive validity for assessment module was provided previously, the effectiveness of real-time coaching and risk detection ability with this AI system remains to be explored.

The objective of this study is to compare the error-oriented intelligent feedback provided by the ICEMS to in-person expert instruction in surgical simulation training by monitoring the improvement of medical student technical skills on a series of virtual reality tumor resection tasks.

Condition or Disease Intervention/Treatment Phase
  • Behavioral: Experimental: Experimental Group - Intelligent Continuous Expertise Monitoring System group
  • Behavioral: Experimental Group In-person expert-mediated instruction group
N/A

Detailed Description

Background Advancements in the fields of surgical simulation and artificial intelligence have resulted in the development of intelligent technical skill assessment and tutoring systems to meet the needs of the competency-based approach in medical education. These novel educational tools offer solutions for standardization and objectivity of the learner's technical competence assessment and training, in risk-free simulated operative environments. These systems can supplement the gold-standard apprenticeship model in training in surgery and other procedural based medicine by providing detailed information about technical skills, action-oriented procedural guidance, and risk assessment. Such applications can help educators in developing objective formative and summative student assessment and teaching methodologies.

Rationale The Intelligent Continuous Monitoring System (ICEMS), an artificial intelligence (AI) deep learning application was developed to assess surgical performance continuously in 0.2 second intervals with the ability of coaching trainees to expert level and risk detection. This system involved in a patent pending entitled 'Methods and systems for continuous monitoring of task performance on virtual simulators' (2020; patent No. 05001770-883USPR). Previously, a predictive validity for assessment module was provided however, the effectiveness of this system in real-time coaching trainees and risk detection needs to be explored.

Research Objectives To compare the effectiveness of an intelligent real-time coaching system to the gold standard expert mediated feedback. To test the transfer of learning during the practice tasks to the more complex realistic scenario.

Hypotheses

  1. The ICEMS feedback group will significantly improve in the composite-score between the first attempt to the last attempt in practice scenario.

  2. The control Group with no-expert mediated post hoc benchmark feedback will not statistically improve in the composite-score between first attempt to the last attempt in practice scenario.

  3. The ICEMS feedback group will have a statistically higher composite-score compared to the control group, in the fifth attempt in the practice scenario.

  4. The ICEMS feedback group will have a composite-score statistically non-inferior compared to the expert mediated instruction group in the fifth attempt on the practice scenario.

  5. The global OSATS score of the ICEMS feedback group will be non-inferior to that of the expert mediated instruction group in the complex realistic scenario.

  6. That ICEMS feedback will not elect a difference in emotion or cognitive load compared to the expert mediated instruction group.

Specific Aims: 1) To assess if the efficacy of AI mediated real-time tailored feedback is statistically non-inferior to expert mediated in-person feedback in improving medical students' bimanual surgical skills on two virtually simulated surgery tasks.

2)To outline if different emotions and cognitive load are elicited by the ICEMS feedback system as compared to human instruction

Design: A three-arm single blinded randomized controlled trial of AI training versus expert mediated training versus no-expert mediated training.

Setting: Neurosurgical Simulation and Artificial Intelligence Learning Centre, Montreal Neurological Institute.

Participants: Students who are enrolled in their premed year or first to fourth year in medical schools across the province of Quebec.

Task: Complete removal of a simulated practice tumor five times and a complex simulated brain tumor once with minimal bleeding and damage to surrounding simulated healthy brain structures using two surgical instruments (Ultrasonic Aspirator and Bipolar pincers) using the NeuroVR (CAE Healthcare) surgical simulator.

Intervention: A single 90-minute training session, including six virtual subpial tumor resection attempts (five simple practice scenarios and one complex realistic scenario) with assessment and feedback from either:

  1. the Intelligent Continuous Expertise Monitoring System (ICEMS) (Group 1)

  2. an in-person expert instruction feedback (Group 2)

  3. no-expert mediated post hoc benchmark feedback (Group 3)

Main Outcomes and Measures:

Primary outcome is surgical performance improvement in the composite-score assessed by the Intelligent Continuous Expertise Monitoring System's (ICEMS) previously validated assessment module. Performance improvement is measured as the composite-score difference between participant's attempts during the five practice attempts. Learning transfer will be assessed by the composite-score obtained in the complex realistic scenario.

Secondary outcome is participants' performance score on the realistic task which will be scored by blinded experts using the Objective Structured Assessment of Technical Skills (OSATS) global rating scale. Differences in strength of emotions elicited will be measured by Duffy's Medical Emotional Scale (MES), before, and after the intervention and difference in cognitive load will be measured using Leppink's Cognitive Load Index (CLI) after the intervention

This study, to the knowledge of the investigators, is the first application comparing the effectiveness of a real-time intelligent feedback system to the gold standard human instruction. This study aims to identify efficient training methodologies using artificial intelligence powered surgical simulator training platforms, hence trainees are provided objective, tailored technical skill assessment and coaching using state of art technologies.

Statistical Analysis Plan Participant data will be anonymized and stored. The Intelligent Continuous Expertise Monitoring System (ICEMS) will be used to score surgical performance using raw performance data and provide a performance score at 0.2 second intervals. An average composite-score will be calculated for each task. Within and between groups comparisons will be made to assess the learning from first attempt (baseline performance) to the fifth attempt and compare improvement between groups. The composite score on the realistic scenario will be calculated by the ICEMS to assess transfer of learning. Participant score improvement will be examined by repeated measures ANOVA. Considering a potential effect size of 25%, a statistical power of 0.95 and a significance of 0.05 this study needs to recruit 29 participants in each group. Participants' performance score in the fifth practice attempt will be compared between groups using ANOVA to compare learning. The same analysis will be done for the realistic attempt to compare learning transfer between groups.

Performance in the realistic attempt will be rated by blind experts using the OSATS global rating scale, using participant performance video. The global OSATS score will be compared between groups to compare efficacy in learning using ANOVA.

Study Design

Study Type:
Interventional
Actual Enrollment :
98 participants
Allocation:
Randomized
Intervention Model:
Parallel Assignment
Intervention Model Description:
Randomized Control trialRandomized Control trial
Masking:
Double (Participant, Outcomes Assessor)
Masking Description:
Double (Participant and Expert Rater) Participants do not know the performance metrics used in calculation of their final composite-score, only that they will be learning and practicing technical skills used in neurosurgery while receiving feedback from an instructor or an intelligent system, in subpial tumor resection procedures. Experts do not know to which group the video performance they are rating belongs, for the OSATS rating.
Primary Purpose:
Health Services Research
Official Title:
Testing the Efficacy of an Artificial Intelligence Real-Time Coaching System in Comparison to In Person Expert Instruction in Surgical Simulation Training of Medical Students - A Randomized Controlled Trial
Actual Study Start Date :
Jan 5, 2022
Actual Primary Completion Date :
May 3, 2022
Actual Study Completion Date :
May 3, 2022

Arms and Interventions

Arm Intervention/Treatment
No Intervention: Control Group No-expert mediated post hoc benchmark group

30 participants. Individuals receive identical introductory information,same time, to perform, same scenarios as other groups. Students receive their scores on 5 performance metrics compared to expert performance benchmarks. Scores are presented in the 5 minute breaks between tasks. Student goal is to be within the benchmark in all five metrics.

Experimental: Experimental Group - Intelligent Continuous Expertise Monitoring System group

30 Participants. Introductory information provided on simulator and scenario. They perform 5 simple practice subpial tumor resections with 5 minutes per trial. On 6th attempt 13 minutes to perform a complex realistic scenario. During first practice task, participants receive no feedback. For the subsequent 4 practice tasks participants will receive real-time auditory feedback instruction by the intelligent system. After each of the 5 attempts, a student takes a 5-minute break. During each of the 5 breaks the participants will be shown the errors they made during the task by the intelligent system regarding five performance metrics monitored. After seeing each error outline, the participant will be shown a video demonstration to learn how to expertly perform on each performance metric. On their 6th attempt they will perform on the realistic scenario without any feedback given.

Behavioral: Experimental: Experimental Group - Intelligent Continuous Expertise Monitoring System group
During each of the practice task they will receive real-time auditory feedback instructed by the intelligent system. After each attempt, a student takes a 5-minute break. They will be shown the errors they made during the task regarding five performance metrics. After seeing each error, they will be shown video demonstration to learn how to expertly perform at each performance metrics. On their 6th attempt they will perform on the realistic scenario without any feedback given.

Experimental: Experimental Group In-person expert-mediated instruction group

30 Participants. Introductory information provided on simulator and scenario. They perform 5 simple practice subpial tumor resections with 5 minutes per trial. On 6th attempt 13 minutes to perform a complex realistic scenario. During first practice task participants receive no feedback. For the subsequent 4 practice tasks participants receive real-time auditory feedback instruction by in-person expert during the task. After each of the 5 tasks, students takes a 5-minute break. During each of the 5 breaks the in-person expert provides feedback to the participant based on their OSATS score assessment during the previous trial. If the expert feels it is appropriate the expert will demonstrate how to do the specific procedure which has been found to be a concern on the simulator themselves so the participant can understand how to improve their performance. On their 6th attempt they will perform on the realistic scenario without any feedback given.

Behavioral: Experimental Group In-person expert-mediated instruction group
During each of the practice task, students will receive verbal feedback from the expert instructor present in the room. After each task, experts will summarize their performance and outline the errors the student made. Based on the student's performance, expert will demonstrate how to expertly perform the task in the simulation, and how to improve their performance in the next attempt. Students will perform the 6th attempt on the realistic scenario without any instruction given.

Outcome Measures

Primary Outcome Measures

  1. Change in performance [1 Day of Study]

    Performance will be measured using the composite-score assessed by the ICEMS system

  2. Transfer of learning [1 Day of Study]

    Performance on the complex realistic scenario will be evaluated by the composite-score assessed by the ICEMS system.

Secondary Outcome Measures

  1. Objective Structured Assessment of Technical Skills (OSATS) global rating scale [1 Day of Study]

    Performance on the realistic attempt will be rated by blinded experts using the Objective Structured Assessment of Technical Skills (OSATS) global rating scale. Efficacy in learning with real-time intelligent feedback system will be compared to expert-mediated instruction and to no-expert mediated post hoc feedback.

  2. Differences in strength of emotions elicited [1 Day of Study]

    Measured by Duffy's Medical Emotional Scale (MES)

  3. Difference in Cognitive Load [1 Day of study]

    Measured using Leppink's Cognitive Load Index (CLI) after the intervention

Eligibility Criteria

Criteria

Ages Eligible for Study:
18 Years and Older
Sexes Eligible for Study:
All
Accepts Healthy Volunteers:
Yes
Inclusion Criteria:
  • Medical students who are actively enrolled in medical school in any Canadian institution who do not meet the exclusion criteria.

  • Premedical students who are actively enrolled in medical school in any Canadian institution who do not meet the exclusion criteria.

Exclusion Criteria:
  • Participation in previous trials involving the NeuroVR (CAE Healthcare) simulator.

Contacts and Locations

Locations

Site City State Country Postal Code
1 Montreal Neurological Institute and Hospital Montréal Quebec Canada H3A 2B4

Sponsors and Collaborators

  • McGill University

Investigators

  • Principal Investigator: Rolando Del Maestro, MD, McGill

Study Documents (Full-Text)

None provided.

More Information

Publications

Responsible Party:
Rolando Del Maestro, Director, Neurosurgical Simulation and Artificial Intelligence Learning Centre, McGill University
ClinicalTrials.gov Identifier:
NCT05168150
Other Study ID Numbers:
  • 2010-270, NEU-09-042-Trial 3
First Posted:
Dec 23, 2021
Last Update Posted:
Aug 8, 2022
Last Verified:
Aug 1, 2022
Individual Participant Data (IPD) Sharing Statement:
Yes
Plan to Share IPD:
Yes
Studies a U.S. FDA-regulated Drug Product:
No
Studies a U.S. FDA-regulated Device Product:
No
Keywords provided by Rolando Del Maestro, Director, Neurosurgical Simulation and Artificial Intelligence Learning Centre, McGill University

Study Results

No Results Posted as of Aug 8, 2022