120 likes | 285 Views
JAEC Assessment Initiatives and Implications. Presented to: NDIA’s Training Transformation Technologies Conference September 5, 2003. Julia Loughran ThoughtLink, Inc. loughran@thoughtlink.com 703-281-5694. 1. Enhanced JTS. JTS. 2. Joint Performance Measures. 3.
E N D
JAEC Assessment Initiatives and Implications Presented to: NDIA’s Training Transformation Technologies Conference September 5, 2003 Julia Loughran ThoughtLink, Inc. loughran@thoughtlink.com 703-281-5694
1 Enhanced JTS JTS 2 Joint Performance Measures 3 Performance Assessment Architecture JKDDC JNTC 4 Joint Lessons Learned Program Linking Training to Readiness 5 JAEC Assessment Initiatives
1 Enhanced JTS JTS • Strengthen and broaden the focus of the JTS • “… by using the JTS methodology as the foundation for service, Federal agency, intergovernmental, and multinational training programs” • Link combatant commander requirements with performance measures • Use this info to guide T2 assessments and investments • Implications for JAEC: • Need to identify current JTS metrics, processes, AAR and lessons learned capabilities (establish baseline) • Broaden the existing scope • Issue of acceptance outside DoD • Consider other approaches (outside traditional education and training) • e.g., COPs, web portal, distributed trainings • Example of JIACGs; interagency culture issues
2 Joint Performance Measures • Initial metrics based on existing tasks, conditions, and measures in the UJTL (“predictable capabilities”) • Second level metrics focus on “unforeseen capabilities” (e.g., cognitive agility, innovative abilities) • Implications for JAEC: • Definition of metrics, required data elements, standardization of assessment techniques • Identify issues related to data standardization, data interoperability • Need for innovative ways to measure typically hard to define characteristics (SSA, agility, adaptability, leadership) • Importance of developing measures for trainers
3 Performance Assessment Architecture JKDDC JNTC • Architecture will help guide the integration of training simulations, instrumentation ranges, and distributed learning systems • Incorporate common assessment tools and metrics in its operational software and global information environment • JAEC Implications: • Need for new automated performance assessment tools/analysis techniques (e.g, PDAs) • Verify/validate metrics for overall system effectiveness • E.g., strategy, technologies, trainers, teams, and individuals • Establish performance assessment repositories
Enhance the Defense Readiness Reporting System • Develop more accurate/insightful methods for measuring the effectiveness of Joint education and training on performance of joint ops • Provide more robust reporting in terms of DOTMLPF (both predictable requirements and unforeseen exigencies) • JAEC Implications: • Provide bridge between individual and collective measures • Need to link home station training to joint training events • Need to define readiness standards across the full range of missions • Enable performance trends analyses 4 Linking Training to Readiness
Joint Lessons Learned Program 5 • Integrate lessons learned into the development of new training processes and systems • Systematically capture lessons learned from operational missions and inject into experimentation and concept development • Build on the Joint Center for Lessons Learned • JAEC Implications: • Need to more tightly couple the training community and the experimentation/concept development community • Highlight ways to improve lessons learned dissemination – focus on lessons learned vs. lessons lost
Other Important Areas to Consider • Communicating the vision • Understanding roles • Breaking down boundaries • Dealing with resistance • Timing • Building trust • Continuity of direction vs. accomplishments • Cultural issues • Addressing fears • Cohesive top management • Investment implications • Preparing the system for change