1 / 32

Shore-Based Squadron Course

Shore-Based Squadron Course. VP, VQ(T), VQ(E), VPU, VR, and VX. 30 June 2014. Introduction. This presentation covers the following topics:. Shore-Based Squadron Course. Background Metrics Overview Assessments Reporting Rules DRRS-N Accounts CNAF Readiness Brief Training

eavan
Download Presentation

Shore-Based Squadron Course

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Shore-Based Squadron Course VP, VQ(T), VQ(E), VPU, VR, and VX 30 June 2014

  2. Introduction This presentation covers the following topics: Shore-Based Squadron Course Background Metrics Overview Assessments Reporting Rules DRRS-N Accounts CNAF Readiness Brief Training Keys to Success

  3. Background – DRRS Background • DRRS provides a single, standardized DoD readiness reporting system based on Mission Essential Tasks (METs) • DRRS improves the assessment, visibility, and force management of our forces/organizations • DRRS is used by all services to report readiness to DoD and Joint Staff and is commonly referred to as DRRS-S (DRRS Strategic)

  4. Background – Mission Essential Tasks • METs are tasks considered essential to accomplish and support missions and requirements assigned by a Joint or Naval Commander • The Mission Essential Task List (METL) is based on the Unit’s ROC/POE and maps tasks to capability areas • Capability Areas = Warfare Mission Areas from the ROC/POE • Squadrons only report on Primary Mission Areas from their ROC/POE • Each MET contains a set of conditions and standards • The METL is common to all units within a Responsible Organization (RespOrg)

  5. Background – DRRS-N • DRRS-N is the tool Navy organizations use to report MET readiness data to DRRS • DRRS-N has two software applications • DRRS-N Ashore (Shore-based) • DRRS-N Afloat (Ship-based ) • Navy Policy Guidance • NTRP 1-03.5 • OPNAV 3501.360 • CNAP/CNAL 3500.38 Note: All references are on the CNAF Readiness Reference Tool.

  6. Background – DRRS-N Ashore Sample

  7. Background – DRRS-N Assessments • DRRS-N contains two assessments: • Computed Assessment (Automated) • PESTO Pillars (Per, Eqp, Sup, Tng, Ord) • Pre-calculated data from authoritative data sources • Squadrons will see only the Personnel, Equipment and Training pillars • Commander’s Assessments (Subjective) • The Commander shall assess Core, Capability Areas, all METs, and the legacy SORTSREPNV C Rates • PESTO Pillar data is provided to assist the Commander in making an assessment • The Commander’s Core, Capability Area, and MET assessments are the only data that flow up beyond Navy lifelines to DRRS Personnel Equipment Training

  8. Background – DRRS-N Data Flow DRRS-N is comprised of a network of interdependent systems, architectures, databases, and applications. The network provides the means for reporting readiness data when Ashore or Afloat. The Afloat systems only apply to ship-based units.

  9. Background – DRRS-N Data Sources DRRS-N compiles unit resource values (called FOMs or Figures of Merit), using authoritative data sources. These values and associated data are available to the Commander as a guide for making unit assessments.

  10. Metrics Overview – PESTO Metrics PESTO pillar data is dependent on inputs, data sources, frequencies, and metrics. The illustration below summarizes the inputs, data sources, frequencies, and metrics for personnel (P), equipment (E), and training (T).

  11. Assessments – Types • The Commander assesses the following areas with a Y, Q, or N: • The MET • The Capability Area • The Core • In addition, the Commander shall update the OARS (Organization and Resource Status) data. OARS will be discussed later.

  12. Assessments – Values The following values are used to assess the Core assessment, Capability Areas, and METs: Commander’s Assessments

  13. Assessments – Values The following values are used to assess the Core assessment, Capability Areas, and METs: Y – Yes (Green) The unit can successfully perform an assigned Capability or MET to the prescribed standards and conditions. A “Yes” assessment should reflect demonstrated performance in training and operations.

  14. Assessments – Values The following values are used to assess the Core assessment, Capability Areas, and METs: Q – Qualified Yes (Yellow) (Used when data may not readily support a “Yes” assessment). The unit is expected to accomplish an assigned Capability or MET to the prescribed standards, under most conditions. This performance, however, has not been observed or demonstrated in operations or training. Units with this rating can be employed for those tasks.

  15. Assessments – Values The following values are used to assess the Core assessment, Capability Areas, and METs: N – No (Red) The unit cannot perform an assigned Capability or MET to the prescribed standards and conditions at this time. Supporting explanations are mandatory for any “N” assessment.

  16. Assessments – TYCOM Guidance • Core assessment shall not be assessed as green if either MOB or CCC are assessed less than green • The Commander’s assessment shall not be green for any Capability Area if the unit fails to complete the full training requirements for any phase of the FRTP • Always Assess to the MCO standard and keep comments pointed and operationally focused • “MCO ready” is defined as a unit’s ability to deliver it’s full design capability as delineated in the ROC/POE and should not be measured against the requirement of a specific OPLAN • Unit Commanding officers should not report “Y/Green” across the board because their unit is green for it’s current FRTP phase

  17. Assessments – Commander’s Remarks • The Core assessment block contains specific sections for your comments: • Overall Section: Describe the unit’s current readiness in relation to the wartime mission and ability to conduct current tasking • PESTO Sections: Describe the specific readiness issues and operational impacts • Comments should focus on current and future capability and issues • Projections on readiness changes should be based on events and not just the next DRRS-N assessment • Minimum comments should include; • Current FRTP (R+) month/phase • Next major FRTP milestone • Expected deployment date • Commander’s top readiness concerns Note: Since only the Commander’s assessment and comments go up to DRRS, remarks about PESTO pillar data should be submitted to CNAF N40 via other means.

  18. Assessments – Commander’s Remarks • Personnel Remarks • Squadron – Comment on IAs, TADs, and other issues that cause personnel shortages • Equipment Remarks • Comment on long term aircraft and equipment shortages • Training Remarks • Use the SHARP skills report to highlight skills that are keeping your TFOM scores low • Pf data does not appear in the T Pillar drill down, so you must have good comments • General Remarks/Tips • Use plain language; this is not a message, so use sentence case • DRRS-N is about MCO capability and not what your unit is tasked with currently • Do not use SORTS terminology in DRRS-N comments • Don’t just write, “Readiness as per FRTP phase” • There is no need to delay an assessment due to PESTO pillars that are not populated or erroneous pillar data

  19. Assessments – Commander’s Remarks The SHARP Crew METs w/Individual Skills report is a useful tool for writing Training comments. This report helps the squadron determine the status of skilled crews (Pf) for any given NTA

  20. Assessments – SORTSREPNV Data Reporting • In addition to the MET assessments, the Commander shall update the OARS data. • CJCS has not released the component services from their responsibility to report SORTSREPNV data • USFF includes the OARS data in DRRS-N to capture legacy SORTSREPNV data • OARS transmits SORTSREPNV data along with the DRRS-N assessment and creates a SORTSREPNV report internal to the larger DRRS network for use at Joint Staff

  21. Assessments – OARS Data Calculations Everyone with a DRRS-N account can view your OARS data. To update the data in the OARS page, select the Edit buttons or by click on the C Rate boxes. The data in the OARS page is filled in by selecting the Edit buttons or by clicking on the C Rate boxes (in the InStaffing view only).

  22. Assessments – OARS Data Calculations • When the user selects either Edit buttons or clicks on the C rate boxes, a data entry • screen will appear. Complete the data fields required. • Report current employment via the Activity Code sets • Report deployment status using the pull-down menu • C Rates should be consistent with the commander’s assessed (DRRS-N) values • Ordnance C Rates for squadrons are N/A • The Supply C Rate is manually calculated based on the PAA of the squadron

  23. Assessments – OARS Data Calculations • Embarked is left blank. In the LAT/LONG report your homeport only • Category Level and Limitations are N/A for CNAF units • Percent Effective is your readiness to perform your currently assigned mission (1-4) • Chem-Bio is normally C4 for squadrons until deployed • Report BA/NMP/COB for all officers/enlisted in the Personnel Strength boxes Note: The guidance for completing the OARS page is contained in Section 5 of the CNAP/CNAL 3500.38

  24. Reporting Rules - Frequency • Submit assessments in DRRS-N as follows: • Within 24 hours of a significant change in readiness • Within 30 days of the last assessment • Following completion of major FRTP milestones to include ARP or Wing directed events • When otherwise directed Note: Your ISIC may also direct a DRRS-N assessment based on COCOM or Fleet Commander reporting requirements.

  25. Reporting Rules – Readiness Expectations • The Fleet Response Training Plan (FRTP) standards are designed to incrementally prepare a unit for its full wartime readiness • During an FRTP cycle, the expectations of readiness increase progressively from “No”(Red)in the maintenancephase to “Yes”(Green) at the beginning of a unit’s employability window • The commander’s assessment is expected to change from red to yellow as the unit moves from the basic phase to the intermediate phase assuming that the unit is on track and is resourced appropriately in the FRTP • Units should strive to maintain a “Yes”(Green) throughout their employability window unless their mission forces a degrade in certain Capability Areas (single mission focus, non-wartime tasking) • Not all units follow this model (HM, VQ(T), VPU, and FDNF) Maintenance Basic Intermediate Employability DRRS-N Note: Currently, MOB is the only Capability Area that is intended to be green prior to the employability phase

  26. DRRS-N Accounts • DRRS-N Ashore • Requested via the SIPRNET (https://drrsn.ffc.navy.smil.mil/drrsn) • Accounts are managed by USFFC • At least, CO, XO, and OPS should have Commander permissions • Help Desk: 757-836-6365 (DSN 836) • Email: drrsn@navy.mil • Tips • Allow 1-2 hours to get a DRRS-N Ashore account • DRRS-N accounts are tied to the individual thus no group access accounts are permitted

  27. DRRS-N Accounts - Permissions The table below describes the DRRS-N account permission levels. CNAF recommends that the CO/OIC, XO, and OPS-O all have the Commander permission level. The CO has the discretion to assign any permission level to personnel in their command. DRRS-N Ashore URL (SIPRNET): https://drrsn.ffc.navy.smil.mil/drrsn

  28. CNAF Readiness Brief Both CNAP and CNAL present a bi-monthly readiness brief to CNAL N00 using DRRS-N readiness data. The brief pages are organized by FRTP phase and deployment groups. Squadrons in maintenance phase are not displayed on the brief. Currently VR and VX squadrons are not included in this brief. Reserve squadrons are added to the brief if they deploy in support of a COCOM or Fleet Commander requirement. This graphic shows a sample VP (deployed) slide. Other slides would display VP (Sustain), RECCE, and TACAMO units. UNCLASSIFIED Sample

  29. Training Plan – Methods Three avenues for DRRS-N training are: • On-line and embedded training • CNAF Readiness Reference Tool • Help functions in DRRS-N • Pipeline training • PXO Course • OIC Courses • ARP DRRS-N Training • Waterfront briefs • Quarterly at Fleet Concentration sites • As requested by units

  30. Keys to Success • CO gets 100% of the vote in the Commander’s assessment regardless of the PESTO pillars • The unit needs to understand their FRTP cycle and their readiness expectations • Remarks should be focused on readiness issues for your squadron/detachment • Do not use SORTS terminology in remarks • Ensure that all remark fields are continually updated • Concentrate your remarks in the Core Comment block • Evaluate all METs, Capability Areas, Core, and OARS on every assessment • OARS data should match the current DRRS-N Capability Area assessments • Manage the input systems (AMSRR, SHARP, etc.) • Flight logging in SHARP really matters • The AMSRR must contains numbers for all aircraft and equipment fields, even if the number is zero • Take advantage of the TYCOM and Type Wing experts

  31. CNAF DRRS-N Pillar Leads CDR Nate “Dude” Ogle DRRS-N PM 757-322-3325 CDR Robert PalmoreE Pillar757-445-6689 Mr John OlanowskiP Pillar 757-444-2640 CDR Timothy “Spock” LeonardT Pillar 619-767-7745 Contractor Support Leads Mr Chris SolerTraining & Policy Development 757-322-3327 Mr Ray SevidalNMETL Management619-767-1191 Mr Dan ScholtesLead SHARP Support Rep757-270-0955 Mr Robert CarlsonLead ADW Support Rep 619-955-5800

  32. QUESTIONS?

More Related