中文(Chinese)
评估工具箱(Evaluation Toolkit)是 IAMM 为全球医疗机构、教育者、研究人员与技术团队设计的标准化评价体系集合,用于评估元宇宙医学(Metaverse Medicine)、AI 医疗、沉浸式训练等多种应用场景的有效性、安全性与可用性。
该工具箱提供多层级、多维度的评估方法,确保学习成果、临床技能、系统性能与用户体验均可量化、可追踪、可比较。
工具箱核心目标:
- 支持全球一致的评估标准
- 为医院与教育机构提供即用型工具
- 促进监管审核、科研设计与真实世界数据分析
- 提升沉浸式医疗教育与 AI 医疗应用的质量保证水平
(1)学习与能力评估模块(Learning & Competency Assessments)
用于教育培训与技能提升,包括:
- 学习前后测(Pre/Post Tests)
- OSCE 客观结构化临床技能考核
- XR 技能评分量表(XR Skill Scales)
- 案例推演评估(Scenario Assessment Rubrics)
- 技能清单(Competency Checklists)
- 学习行为分析(Learning Analytics)
(2)临床效果与安全监测(Clinical Outcomes & Safety Monitoring)
适用于元宇宙临床场景、虚拟诊疗、vICU/vMDT、数字疗法等:
- 症状与功能改善指标(Symptom/Function Outcomes)
- 患者依从性(Adherence Metrics)
- 不良事件(Adverse Events)
- 信息与技术错误记录(Error Logs)
- 远程操作风险评估(Remote Operation Safety Metrics)
(3)沉浸感与体验质量评估(Immersion & Experience Evaluation)
针对 XR / 虚拟系统的核心指标:
- 沉浸感量表(Presence & Immersion Scales)
- 用户满意度(Satisfaction Surveys)
- 认知负荷(Cognitive Load Measures)
- 舒适度与晕动症评估(Comfort & Cybersickness)
- 可用性(Usability Scores)
(4)系统性能与可靠性工具(System Performance & Reliability Tools)
用于测试 XR、AI、大模型、数字孪生等系统:
- 网络延迟(Latency Metrics)
- 渲染与同步性能(Rendering & Sync Tests)
- 系统稳定性(System Uptime)
- 隐私与数据安全检测(Privacy & Security Checks)
- 互操作性测试(Interoperability Tests:FHIR/DICOM/XR APIs)
(5)科研设计与数据分析组件(Research & Data Components)
支持多中心研究与真实世界数据(RWD)分析:
- 结局指标选择指南(Outcome Measurement Guides)
- 样本量估算工具(Sample Size Estimators)
- RCT 与 RWS 模板(Trial Design Templates)
- 数据质量控制(Data Quality Checklists)
- 统计与可视化工具包(Stats & Visualization Toolkit)
(6)监管与伦理支持材料(Regulatory & Ethical Support)
用于医院、科室、伦理委员会、企业:
- 风险评估模板(Risk Assessment Forms)
- 伦理审查要点(IRB Key Considerations)
- 数据保护与同意文件(Consent & Data Protection Forms)
- 合规检查表(Compliance Checklists)
English(English Version)
The Evaluation Toolkit is IAMM’s comprehensive standards-based assessment suite, designed for healthcare institutions, educators, researchers and technology developers to evaluate the effectiveness, safety and usability of Metaverse Medicine, AI-driven care and immersive training systems.
Its multi-level, multi-domain components ensure that learning outcomes, clinical performance, system reliability and user experience are measurable and comparable across global settings.
Toolkit objectives:
- Support globally aligned evaluation standards
- Provide ready-to-use tools for hospitals and academic programs
- Facilitate regulatory review, research design and real-world data studies
- Enhance quality assurance for immersive and AI-enabled medical applications
(1) Learning & Competency Assessments
For education and skills development:
- Pre/Post tests
- OSCE (Objective Structured Clinical Examination)
- XR skill rating scales
- Scenario-based assessment rubrics
- Competency checklists
- Learning-analytics dashboards
(2) Clinical Outcomes & Safety Monitoring
Applicable to virtual care, vICU/vMDT and digital therapeutics:
- Symptom/function outcomes
- Adherence metrics
- Adverse-event tracking
- Technical error logs
- Remote-operation safety metrics
(3) Immersion & Experience Evaluation
For XR/virtual systems:
- Presence & immersion scales
- Satisfaction surveys
- Cognitive load measures
- Comfort & cybersickness scales
- Usability scores
(4) System Performance & Reliability Tools
For testing XR, AI, FM models and digital twins:
- Latency measurements
- Rendering & synchronization performance
- System uptime
- Privacy & security audits
- Interoperability testing (FHIR, DICOM, XR APIs)
(5) Research & Data Components
Supporting multi-center research and RWD analytics:
- Outcome-measurement guides
- Sample-size calculators
- RCT/RWS design templates
- Data quality checklists
- Statistical & visualization toolkits
(6) Regulatory & Ethical Support Materials
For hospitals, ethics boards and industry partners:
- Risk-assessment forms
- IRB key review items
- Consent & data-protection templates
- Compliance checklists