270 likes | 587 Views
One Semi-Automated Force (OneSAF). User Engagement Points and Processes Session 6.1 Doug Parsons Chief Engineer, OneSAF Software Development. OneSAF processes overview Help desk User Forum User feedback Change Request/PTR processes Build process External developer handover process.
E N D
One Semi-Automated Force (OneSAF) User Engagement Points and ProcessesSession 6.1Doug ParsonsChief Engineer, OneSAF Software Development
OneSAF processes overview Help desk User Forum User feedback Change Request/PTR processes Build process External developer handover process User Processes & Engagement PointsAgenda Briefing objective: Discuss OneSAF processes that engage with our users and inform users of entry points into the OneSAF system.
User Processes & Engagement Points Process overview • The Architecture & Integration (A&I) team was tasked from program initiation to develop a robust set of processes that would guide development of the OneSAF system. • Unlike typical software acquisition programs PM OneSAF is essentially the ‘prime’. • Developing a set of processes that could guide the efforts of a variety of different task orders became essential. • These process have been certified at SEI CMMI level V (July 2004) and ISO 9001:2000 (June 2007). * * CMMI audit was performed on SAIC's Integrated Mission and Simulation Systems Operation (of which includes OneSAF A&I).
User Processes & Engagement Points Internal processes • The intent is not to provide a review of our internal software development processes or practices; however, there is value in pointing you to where these processes can be found. • https://dev.onesaf.net/Architecture_and_Integration/index.html • Pay particular attention to the Electronic Process Guide (EPG) • Knowledge Engineering and Conceptual Modeling • System definition • Product line definition • Component development • Integration & Test • Internal and external handover process
User Processes & Engagement Points Electronic Process Guide 1. From the OneSAF.net internal development site (https://dev.onesaf.net/ or through the OneSAF Community page) find the Architecture & Integration Task Order link. 2. Locate the EPG button as shown. Note that some of these processes are being reviewed and revised to more appropriately reflect post OneSAF v1 development and release activities.
User Processes & Engagement Points Help Desk Process • Addresses a wide variety of issues and ‘how do I’ questions. • Installation ▪ Databases • Training ▪ Operation • Scenarios ▪ Exercises • Model development ▪ Knowledge acquisition • Knowledge engineering ▪ System administration • Software development • Most prevalent questions addressed to date: • Installation/media – 13% • OneSAF.net – 10% • Terrain database - 8% • General info/data requests – 7% • Tools operation (MCT, SCAMT, MSDE, etc.) – 7% • DIS and HLA – 6% • Behavior models – 5% • Network/distributed operations – 4%
User Processes & Engagement PointsGetting help desk support Links to help desk support
User Processes & Engagement Points User Forum • Purpose is to facilitate and foster cross-OneSAF community discussion and collaboration. • Tips • Problem resolution • Good ideas for future capabilities • Sharing plans for co-development and/or future use (i.e. exercise or experiment). Discussion boards (most of them): Technical Support • Installation and Configuration ▪ General operation Scenario Execution and Generation • Scenario creation ▪ Terrain • C2 ▪ AAR • Multi-Node, Interop and Performance Analysis • Repeatability and Replication ▪ Data Loading/Collection • Model verification Development and Maintenance • Compositions ▪ ERC/SNE • Infrastructure ▪ KA/KE • Tool, Agent and Modeling Services ▪ System services
User Processes & Engagement Points User Forum
User Processes & Engagement Points User Feedback • Where the help desk will address a wide variety of questions, the user feedback tool offers a mechanism to describe problems or make requests for new capabilities. • Help desk personnel will forward appropriate help tickets to the user feedback tool. • User feedback tickets are regularly reviewed by Gov’t review board. The User Feedback Review Board (UFRB) will appropriately disposition these tickets: • Problem Trouble Report → Engineering CCB • Enhancement request → TPO → Requirements process • Return for more information • Cancel
User Processes & Engagement Points Entering User Feedback Tickets
User Processes & Engagement Points Good User Feedback (1/2) Purpose is to help us duplicate the problem and distinguish symptom from root cause. • Descriptive title of user feedback. • Standalone or distributed? If distributed, • Compositions running in exercise • Number of machines; number that are MCTs vs. Simcores • Number of entities • Have any nodes been late-joined? • Interoperating with DIS? • Which other applications are in use? • Interoperating with HLA? • Which federates are involved? • Which FOM and RTI are you using? • Interoperating with C2 systems? • Which messages are you using? • Which devices/versions are you using? • Terrain database. • Issues involving units/entity should include composition name to include resolution. • Order(s) issued to composition. • Was a work-around used? • Attaching a log file or scenario file capturing the problem is extremely helpful.
User Processes & Engagement Points Good User Feedback (1/2) • Observation • Detailed description of the issue, such as: • Location of units/entities when the observation was made (e.g. 10 digit grid coordinate). • What was the OneSAF login used? • Item selected on editor, terrain features, etc. • What you expected vs. what happened • What units/entities were executing when the observation was made • What was the problem? Good user feedback description, as well as this sample, can be found on OneSAF.net community site by selecting Documentation → Processes → User Feedback and then finding User Feedback Tool on the page.
User Processes & Engagement PointsPTR Process https://www.onesaf.net/community/systemdocuments/Documentation/ProcessGuide/ PTR/PTR_process.html UFRB
User Processes & Engagement PointsChange Request Process UFRB
Peer Reviews Integration of TPO/domain participation into the development cycle. Lesson-learned – Don’t close key peer reviews without direct engagement (i.e. face-to-face or telecon) with user. Being employed successfully with the Knowledge Engineering/ Conceptual Modeling process. Dependency meetings Coordination with internal task order activities Coordinate handovers from co-developers Peer review participation Discuss PTRs for upcoming release Schedule ECCB and PRB meeting Engineering CCB Conduct technical assessment of integrations impacting the baseline (e.g., hardware/software baseline changes, external integration). Make recommendation to the CCB regarding content of the release baseline (i.e., v1.1, v1.5, v2.0…). User Processes & Engagement PointsBuild process
User Processes & Engagement PointsProcess Flow for Future Builds Handover to I&T (for next build) Integration complete Test complete Dependency mtg (for next build) Build kick-off 1 2 3 4 5 6 7 8 9 10 Ten week build cycle • During each build three simultaneous activities are ongoing. • Knowledge engineering • Code and unit test • Integration & Test • Dependency Meetings • Task Order and co-developer schedule • Handovers from co-developers • Peer review participation • Schedule UFRB meetings • ECCB Meetings (Agenda)
User Processes & Engagement PointsBuild 5 Important Dates • Build 5 Start – 16th July 2007 • Build 6 Dependency Meetings • Internal - 28th August 2007 (800 – 1500) • External - 30th August 2007 (1400 – 1700) • Build 5 Complete – 21st September 2007 • Build 6 Start – 24th September 2007 • Build 6 Complete – 30th November 2007
User Processes & Engagement PointsSoftware Handover Process Architecture Compliance Process
User Processes & Engagement PointsSoftware Handover Forms • Signature and Feedback Form • This form documents completion at each phase of the handover process. • External Development Handover Form • This is the biggie! This form is where you begin in putting together a handover package. It contains all the documentation required, and criteria necessary, for a smooth and successful handover. • New/Modified Files Change List • This list should correspond with the files contained in the source build (see Software Included on the External Development Handover Form). Please include full paths for all files. • Known Handover Issues • Pretty self explanatory. You got "uh oh's" coming in? Let us know about them! • Analysis/Design Document • What did you add or change, and how did you accomplish the addition/modification? Feel free to add additional pages or make the cells bigger to accommodate your design descriptions. • List of Test Plans • List any test plans that were developed to test your new and/or modified functionality. • List of JUnit Tests • List all JUnit Tests that are new or were modified with your handover. • Domain Documentation List • Did you add new units? new models? new data? Excellent! This Document allows OneSAF to maintain traceability on the source of the new "stuff". Where did you get that data? What is the justification for those models and units?
User Processes & Engagement PointsCompliance Assessment Checklist External developer performs self-assessment A&I performs code review Infrastructure Compliance Platform Compliance User Interface Compliance Process Compliance Deployment Compliance Verification Compliance Performance Compliance
User Processes & Engagement PointsIssues and Consideration • Lessons learned • OOSC • USMC CACCTUS • SMDC • CGA • Two potentially opposing conditions: • Need to keep the handover process relatively painless for the external developer. Otherwise, they may not be willing to provide code back for integration. • Need to assure that the external developer provides the appropriate review and documentation. Otherwise, quality of the baseline may suffer and contain support gaps for future users. • We are currently working with the A&I team to improve the handover process and identify automated tools.
Questions and Discussion User Processes & Engagement Points