Upload
iola-wilcox
View
41
Download
0
Embed Size (px)
DESCRIPTION
JAEC Assessment Initiatives and Implications. Presented to: NDIA’s Training Transformation Technologies Conference September 5, 2003. Julia Loughran ThoughtLink, Inc. [email protected] 703-281-5694. 1. Enhanced JTS. JTS. 2. Joint Performance Measures. 3. - PowerPoint PPT Presentation
Citation preview
JAEC Assessment Initiatives and Implications
Julia Loughran
ThoughtLink, Inc.
703-281-5694
Presented to:
NDIA’s Training Transformation Technologies Conference
September 5, 2003
JAEC Assessment Initiatives
Joint Lessons LearnedProgram
5
1 Enhanced JTS
JTS
3 Performance Assessment Architecture
JKDDC JNTC
4
Linking Training toReadiness
2Joint Performance Measures
1 Enhanced JTS
JTS
• Strengthen and broaden the focus of the JTS– “… by using the JTS methodology as the foundation for service,
Federal agency, intergovernmental, and multinational training programs”
• Link combatant commander requirements with performance measures – Use this info to guide T2 assessments and investments
• Implications for JAEC: – Need to identify current JTS metrics, processes, AAR and lessons
learned capabilities (establish baseline)– Broaden the existing scope
• Issue of acceptance outside DoD• Consider other approaches
(outside traditional education and training)– e.g., COPs, web portal, distributed trainings
• Example of JIACGs; interagency culture issues
• Initial metrics based on existing tasks, conditions, and measures in the UJTL (“predictable capabilities”)
• Second level metrics focus on “unforeseen capabilities” (e.g., cognitive agility, innovative abilities)
• Implications for JAEC:– Definition of metrics, required data elements, standardization
of assessment techniques– Identify issues related to data standardization, data
interoperability– Need for innovative ways to measure typically hard to define
characteristics (SSA, agility, adaptability, leadership)– Importance of developing measures for trainers
2Joint Performance Measures
• Architecture will help guide the integration of training simulations, instrumentation ranges, and distributed learning systems
• Incorporate common assessment tools and metrics in its operational software and global information environment
• JAEC Implications:– Need for new automated performance assessment tools/analysis
techniques (e.g, PDAs)– Verify/validate metrics for overall system effectiveness
• E.g., strategy, technologies, trainers, teams, and individuals
– Establish performance assessment repositories
3 Performance Assessment Architecture
JKDDC JNTC
• Enhance the Defense Readiness Reporting System– Develop more accurate/insightful
methods for measuring the effectiveness of Joint education and training on performance of joint ops
– Provide more robust reporting in terms of DOTMLPF (both predictable requirements and unforeseen exigencies)
• JAEC Implications:– Provide bridge between individual
and collective measures– Need to link home station training to
joint training events– Need to define readiness standards
across the full range of missions– Enable performance trends analyses
4
Linking Training toReadiness
• Integrate lessons learned into the development of new training processes and systems
• Systematically capture lessons learned from operational missions and inject into experimentation and concept development
• Build on the Joint Center for Lessons Learned• JAEC Implications:
– Need to more tightly couple the training community and the experimentation/concept development community
– Highlight ways to improve lessons learned dissemination – focus on lessons learned vs. lessons lost
Joint Lessons LearnedProgram
5
T2/JAEC’s Success Is Not Just About the Technology!!
Other Important Areas to Consider
• Communicating the vision• Understanding roles• Breaking down boundaries• Dealing with resistance• Timing• Building trust• Continuity of direction vs. accomplishments• Cultural issues• Addressing fears• Cohesive top management• Investment implications• Preparing the system for change